var/home/core/zuul-output/0000755000175000017500000000000015146173023014527 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015146175560015503 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000202272515146175505020271 0ustar corecoreEikubelet.log_o[;r)Br'o-n(!9t%Cs7}g/غIs,r.k9Gfԅ%~6"mv?_eGbuuțx{w7ݭ7֫^|1Fr_?c^*߶E٬:rv筼ح_y~̎+\/_p/Bj^ֻ]Eo^O/(_/V?,<']_kmN:`Si{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~lys%v6:SFA֗f΀QՇ2Kݙ$ӎ;IXN :7sL0x.`6)ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہwħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=d]' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGt?}=ˢ>f>\bN<Ⱦtë{{b2hKNh`0=/9Gɺɔ+'Х[)9^iX,N&+1Id0ֶ|}!oѶvhu|8Qz:^S-7;k>U~H><~5i ˿7^0*]h,*aklVIKS7d'qAWEݰLkS :}%Jk3GIĨmIEQ«` RPZ(D2vh>3fs䓯ҴgqmubIfp$HhtLzܝ6rq/nLN?2Ǒ|;C@,UѩJ:|n^/GSZ;m#Nvd?PqTcLQMhg:F[bTm!V`AqPaPheUJ& z?NwpGj{VjQS,؃I'[y~EQ(S +mpN, Mq 70eP/d bP6k:Rǜ%V1Ȁ Z(Q:IZaP,MI6o ޞ22ݡjR:g?m@ڤB^dh NS߿c9e#C _-XѪ;Ʃ2tStΆ,~Lp`-;uIBqBVlU_~F_+ERz#{)@o\!@q['&&$"THl#d0 %L+`8zOҚƞ`wF~;~pkѽ)'cL@i]<ք6ym®Yi&s`dyMX](^!#h k:U7Uv7쿻чd)wB5v-)s蓍\>S[l52, 5 CۈP$0Zg=+DJ%D  *NpJ֊iTv)vtT̅Rhɇ ќuގ¢6}#LpFD58LQ LvqZDOF_[2arH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{K0D"\KjPQ>Y{Ÿ>14`SČ.HPdp12 (7 _:+$ߗv{wzM$VbήdsOw<}#b[E7imH'Y`;5{$ь'gISzp; AQvDIyHc<槔w w?38v?Lsb s "NDr3\{J KP/ߢ/emPW֦?>Y5p&nr0:9%Ws$Wc0FS=>Qp:!DE5^9-0 R2ڲ]ew۵jI\'iħ1 {\FPG"$$ {+!˨?EP' =@~edF \r!٤ã_e=P1W3c +A)9V ]rVmeK\4? 8'*MTox6[qn2XwK\^-ޖA2U]E_Dm5^"d*MQǜq؈f+C/tfRxeKboc5Iv{K TV}uuyk s" &ﱏҞO/ont~]5\ʅSHwӍq6Ung'!! e#@\YV,4&`-6 E=߶EYE=P?~݆]Ōvton5 lvǫV*k*5]^RFlj]R#Uz |wmTeM kuu8@8/X[1fiMiT+9[ŗ6 BN=rR60#tE#u2k *+e7[YU6Msj$wբh+8kMZY9X\u7Kp:׽ ^҃5M>!6~ö9M( Pnuݮ)`Q6eMӁKzFZf;5IW1i[xU 0FPM]gl}>6sUDO5f p6mD[%ZZvm̓'!n&.TU n$%rIwP(fwnv :Nb=X~ax`;Vw}wvRS1q!z989ep 5w%ZU.]5`s=r&v2FaUM 6/"IiBSpp3n_9>Byݝ0_5bZ8ւ 6{Sf觋-V=Oߖm!6jm3Kx6BDhvzZn8hSlz z6^Q1* _> 8A@>!a:dC<mWu[7-D[9)/*˸PP!j-7BtK|VXnT&eZc~=31mס̈'K^r,W˲vtv|,SԽ[qɑ)6&vד4G&%JLi[? 1A ۥ͟յt9 ",@9 P==s 0py(nWDwpɡ`i?E1Q!:5*6@q\\YWTk sspww0SZ2, uvao=\Sl Uݚu@$Pup՗з҃TXskwqRtYڢLhw KO5C\-&-qQ4Mv8pS俺kCߤ`ZnTV*P,rq<-mOK[[ߢm۽ȑt^?bl\)cb<>O0BďJrDd\TDFMEr~q#i}$y3.*j) qQa% |`bEۈ8S 95JͩA3SX~߃ʟ~㍖›f!OI1R~-6͘!?/Vvot4~6I@GNݖ-m[d<-l9fbn,'eO2sٟ+AWzw A<4 }w"*mj8{ P&Y#ErwHhL2cPr Wҭюky7aXt?2 'so fnHX zַ; ?n` *3UP0Sp8:>m(Zx ,c|!0=0{ P*27ެT|A_mnZ7sDbyT'77J6:ѩ> EKud^5+mn(fnc.^xt4gD638L"!}LpInTeD_1ZrbkI%8zPU:LNTPlI&N:o&2BVb+uxZ`v?7"I8hp A&?a(8E-DHa%LMg2:-ŷX(ǒ>,ݵ𴛾é5Zٵ]z"]òƓVgzEY9[Nj_vZ :jJ2^b_ F w#X6Sho禮<u8.H#',c@V8 iRX &4ڻ8zݽ.7jhvQ:H0Np: qfՋ40oW&&ף \9ys8;ӷL:@۬˨vvn/sc}2N1DDa(kx.L(f"-Da +iP^]OrwY~fwA#ٔ!:*땽Zp!{g4څZtu\1!ѨW(7qZcpL)ύ-G~^rFD+"?_h)yh=x>5ܙQ~O_e琇HBzI7*-Oi* VšPȰһ8hBőa^mX%SHR Fp)$J7A3&ojp/68uK͌iΙINmq&} O L-\ n4f/uc:7k]4p8wWLeUc.)#/udoz$} _3V6UݎvxyRC%ƚq5Щ/ۅw* CVo-1딆~ZYfJ"ou1ϵ5E bQ2mOΏ+w_eaxxOq:ym\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{+ɠ^^fRa6ܮIN ޖ:DMz'rx#~w7U6=S0+ň+[Miw(W6 ]6ȧyԋ4ԙ./_A9B_-Z\PM `iĸ&^Ut (6{\٢K 5XGU/m >6JXa5FA@ q}4BooRe&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jc kkA ~u?u7<?gd iAe1YB siҷ,vm}S|z(N%Wг5=08`S*՟݃*־%NǸ*kb05 V8[l?W]^@G:{N-i bɵFWǙ*+Ss*iނL?{ƱPs(9vb$Η} % 'ןYI ǧPOUCEm<}:H(^jJ=䄸-m!AdEږG)շj#v;#y/hbv BO Iߒ {I7!UՆGIl HƗHg1$ʇȠh#CBΤ{sTQ{%w)7@y1K^ ].Y$46[B-3%OONw8d`Q4d$x0t8@t]y1T\YAidtxBG:pɨyeNg4n]M؞ e}Wn6׳i~'ہZ*FU{fXڃP'Hl4 ,ŸqMHDCYZz Qnz܁$Jp04ȴIL΃.0FiO-qy䶬j顒x.Ƞ"m@Hy_I )j|s#RGI!dTKL&4K>#stV \'xMgaSZNg8>e!^f%cYr]qs:"̊;isXa]d+"v=x7p.fZCg_Ys;pE&\U}ܫSh])qKYAـhΔ`UаQ/ .D%ES*;OLRX[vDb:7a}YF30H #iSpڮޓJ|̖D?:3mhW=rOf'/wѹ8BS8]`;=?Ųm gp8펠ˋֆ:1IC8qٞ\vXçL ]X/r}7O}Wh,h ;RQ=]u10yiC۔I^3!?H6iUV$aZĄH(CR9m*ǁxFd q w帖k8UKxH Hx?tyteIGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dc,m+-!sFɸv'IJ-tH{"KFnLRH+H6Er$igsϦ>QKwҰ]Mfj8dqW+"/fC Q`B 6כy^SL[bJgW^;zA6hrH#< 1= F8) 򃟤lŏd7>WKĉ~b2KQdk6՛tgYͼ#$eooԦ=#&d.09DHN>AK|s:.HDŽ">#%zNEt"tLvfkB|rN`)81 &ӭsēj\4iO,H̎<ߥ諵z/f]v2 0t[U;;+8&b=zwɓJ``FiQg9XʐoHKFϗ;gQZg܉?^HtpR:4Si גt'MH6@w钐֋H{xYEE>6nOj5~IJ|~!yKڮ2 h 3x}~ے4WYr9Ts] AA$ұ}21;qbUwRK #}u'tLi'^Y&,mCM)eu㠥Ѻ\a}1:V1zMzT}R,IA e<%!vĉq|?mtB|C ?dXuWLGml?*uTC̶V`FVY>ECmDnG+UaKtȃbeb筃kݴO~fފ⊈ 8MK?:mM;ߵoz+O~e3݌ƺ(ܸf)*gCQE*pp^~x܃`U'A~E90t~8-2S󹞙nk76s&"mgVKA: X>7QQ-CDC'| #]Y1E-$nP4N0#C'dvܸȯ.vIH"ŐR ;@~{>Kv{) 9AG ćͩ$.!б~N8i"1KФ\L7/,U@.ڮO?mُa ې!rGHwɚ YZgދ8H KV,XHS4OAH$,mr달v9dB.bq:__xW|1=6 R3y^ E#LB ZaZd1,]ןkznxt[|d7W8?XfNѕ|O'  Qk^K|ooz-3,\ϦRY_TBZyX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?nS[{'b{}IStgq$<$ilb)n&  $uT{wD]2cM(%YjDktByxVlᷳ1~jpd1O9Á%. er@3?]^ʧM <mBɃ|.a]]i-*ػWeiبb@66%A,՘|bOc 7O'T*|h^U{ãeeH7nz#<7=li$2cPs6D>TE-&I5&lҬt񗽐eYpCŋ݃k-$pxbڲ&6*9mg>{rtD)wQ`pkKyt1?[ˋڄ5NhfӛŮ Qu8Y4ݯ^+՗~te\g8zHT%{,#/0'NZ׽[q^ė2ymGqF8RŔ.MMWrO»HzC7ݴLLƓxxi2mW4.@`tF)Ċ;@@twml"Ms>\΋"?|NKfֱn !s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O K?md+F/F4is6g],4uWp)ɎT[7jOpH +^X$i,2'=%2VIEwզHe1V%ɖxW̏U r/VcՅ'$0V(D <#JqTؔ4ƪ(pj)V1*^xuQVNWqH.?I!T$2V?DCGQ"h-œ2R,:HWi ƪ7~{1Ǫ zνU%Km|oLTQi^G/iTB( xO4n}sgc@48=2,o{%{%gnr__jy/L|u6W}Ҝp󺸲u ~lw\y+҄WC]s=hc~ۚXj؆턁r?Jt[uo 5YQ͘jj |Y cJa˗Z<0?#u 6A3Y7{<^hgw'U^d{E.W} n"=ß tZXI?oj/EƧ,+|1>55T{&Ph2W1=_WU[[~D; gKSQ҂WxPȐL㻦cp>NuU.UR7/URu/ kj3^O<)SF-(,?b@xDƑ?~Ӿ7u[71B!xDc#֛'3] Uw,H~}yeǸBY~Vچ3k hT&s0_e_g^,ijf wPO1)we tSほW#(P@f~5zjw ),CkYFEV"k G>Bql2 K34Piفfje-32W}eƨqCn%:_O&r}#]-,̋dZ^Ί+qG秧!9I@,gg YGEU8PpXnsVQɟLg⚯!gOh);"6DxrU]GQwה?*E6!ip]>]1O)_I,d]9Nq,F"fnWy_/R6:A\եHT. #PGA!}Z{ Y%F[O%bͿ'|8DQx D͵$8R%$7i. xx)0R;v$ xqE$4&kd):Uv%RZx,:<y\E Ug(M<x'jICr$YP]G5_5i]%~ D# * c{h> ǐ)QМ>}f{z MSU6 @UD>J{4A0$j"Z(H94RoQdˁK>?bVid씲Va-KbRYRjM@ R`:ZOb$s_!p{!hR,vXǴσeTfm+ ]~SfK閰wxF3|Fj3+OSg ( [ DI99aD0Q{PF<9z!\uˑVT C .Q41QR rI{-s:<Ȯ=dВhimMO2]8Mܶ.!/Rlx KG7zCG oj X7֒Ȫїu9 1*2kC7Lp]*xZFsvPy742WKDD^m9|C, FxJ+< `\`NqLϓ9GqZjCT ʖ3Mnv[J n)Q=L&sq0PV|R}jh.%!SY]+!r"ԡ% +fK/xTn7µ$*" fjDG4thݞ`[H]Y>s!|̰Z vȂafj]Y>)%jkVU>ueqM%\JQH<2id],B0oY N#%U{;tU]BϷ)FU]bC%?ڔ@$ x :i$JtmA-`U+ Ծʼ: XvTOP[%jb7Iinv 1ZJY9mxn2>ϹU+%*TwTGj}[GS=&RJmN#0Ŷ/`4F2apEzR.fc;.^rH$\J0 Ad:>F(+.b5?-BPe$o L9_H{KO ?4CFr,dRM6:ʊF#:;nmmH( YbX=3H3v fElb1 6K]'S-VbNS͆T#,h6K<,{5l%{boz[Xt}_(8,~Fal /5&Iz(f/[ĶkIn9!:Y(;<0T^uNzч_jjtkhJWr4XXoe5 t]05=0U]]|zmpfaڡKs{ݍ7ArLٞTK(/pKWX.[OlcnI: (6,#IF eJH}ʡM0G,W[HK[89DK!:n:NO[*^E종ID}Ё%ٲ?zmʖFLv6L==,U4wm.>C>HZRzzxup}În{qx:RTRZYM}6c~F"=xწ ]l$_;C1\"²ڢ :H9띏2:g|kdqRMzFX0#g3 -SM u@:Sh`IlGzFXlSvKL%-Kr[!39mB[f.T,c2\ ,uf6$uUMވ|!Qzܴ 2l滲ⲽ=β+sbqR:j<#Ye$+kHaTj2U*˳8QT'Hs{ȗ~G6iIyU%4 FkK.SeYea":f˃fO5Wֽۏ̰- [f(|>uA1&a9vfh^bn|j T_3-w}Ah8tG6_ş.5_#3tdXƛU0-3iO `=)LBUd"g&Bt(mm(,K?>f7Y&ۓ̲=k==ٙŴWщY_ +1`ܽ1lJ2[7 m+lzߊf|,iŷ *DmAyyf6qgwN|&L)]so0Gn!$^H} [93H" Lœ!$ɜVC9yܸu,c?zq3!@v}эh5v*&Bii΁Z1~4 5cCOHx9_M\gDSp7w*-x?gb1RLAGk@ʐ|n݇N[)VP@&'?~)_UYteqLd9yNny'Tc~Ӎ[ڏw?,өNCw&t_MIVYXŻ '1,d{³KwWjŢ ]y0ʺϿ?p2꾕v7`S6IFKsPw+PCk<&C~x?Yׇ7~| 0{j"(!>°r`'l2Ƕ!\xk]Cz=8QA qc * p7k‚. 9B wO@҂]=Q; ޟcJgQBRTP1x^ܕN7C"zxH}@D]! ~Pwk~8g0qxK;&! [ӛ/8<x8 BIAI D<%[V㫞."@‡ dLCdWD$hO<`A Bz~mĀ  ( k1{+Pũ} L#x q&[D/B0 .zʋ@~Pҳ('B!!( DGE XA1ey]F,"K?Vo1w+(iy/Cr?ڄ_,,>+-\"P&&w:;Dc3^ZwǓ6 g3m8]j!E<)+{s3&>,m( ?[6vM}Qڊ|9^ o鑩'^QɥgAIMiɁi4VyTs nҧaK"} }R>N"3=-<{Z6;\e`F.+@<Цd:_ٰP*="*!UwN*WE-)+=qP^)Y|<0c*EwR,gpSbN:Nb = ؟5lc=(ncqŧHq3gv|;jTǫ[D+*TaE Eh:fGyJXē6>яKƥ9_ο63@WCs0͠HWG$+I>{U`))U9r,dʅ1ږp|hEpX[EI$.J)r,Qqkq-kZz#TOf/-GYA&E~EuTxTͿdXk|W3 [ma}B,΢mrǪr*~hÁ+f)akU yVm4R+-(c3/ 0ς7:ϊ1\ (pCP|pj%\F:0>p t:ɘrWcmrJtJV>ljC-*v<16!oyRr'P,擒a'w!c] .H4_bb(mJzz~xpW!a b(jNgE;vօ)O.&q傭6Ƃ?(/;F1w⾻?P񀻜 |9%yqt/a#?>$Wxv ~ry[J13 ˷zTv^-_b_B(2+/oq|&~>>M[oP aS닓PЪZ^'xb a\$\|TęOY<Ǯ;ar3t xiLDS`ʘx`E 'cF^vʋX2eO" 7ccAiS<0fq;jɗ˰i_C=!);ta(Eew6TfEtm=Lܣn44݃BUeB)_ԻUV>/=[zFm>򼿺 2[̣Wk|êp x*|X>pQgɕ>ՃuV^9d| E#xWe!eE"qBk}%iQQ % m{Q-dB=4/TdeAS7> rj-RYU'r(Q rT'd4 5m-g87'9 WU`3G5QXsG-`[&=]rʫa:Sh*,4IUJJm:e~f6ɋ8vx8;g'::l9ޭ3%)ezt'JHz+}> W6\2|Ѯ7SHfR3}PH]v7)- MZ/# 9C*(Ny sNxeWU_3`hgz[^R U=M ,'x,/tʁQox*B6Ch<UqFkcgپWGouG%_}!}ָbt&eƧ —h_l[hN Q{4 8Tă_ju &8 *jQo9F kYvκ_ #Aql mvuϠH"c-Ftm ;֐f1Lƅ^H;sƳT+ +Jw:nMo,3Xy^$L7axoU\ڑo|e tOlymkNok-`XC0nGGSc^,km+IeA_l`Anه*Ѣ( IKS}HJgĉ]:K,K)|{}uKKnC[Z.5}aW_O!ͽɊK/ǧǐ;?Ui*I3`\8:DD]88CEw ᇿO;mClޡ |Yo4a<^D7jb?FVDZQE~a :}MX߭˿PyW^r*/U왣鯗.:I>z q{}y k/:U}KO ݪ{r/2=ek4LJ1ߘ-g>ts).ѡstɗ_.L\'v9v23Tq{;L)9k3w7R>Ɔ@6o:INֿ*=61#OpS7}j2uf)kFlKpcv/f?`$}J=سgӔD;m0 w$ l 4Mdk ܧȼ)"Í֩3~* n,oMP4FZjpJt 4fKJ'0#%Y(Y?_] W!sfL3-mbQ&ZJٻC_#L]=ODNCl;jf.(Pg:D1QB k@\ot/w0ݵQ-@'Q [S J w1^k3}P"3U=%o-A#օG%UJ_Aebe4`$qy{YhOQ,hp(wNU82XjIGz۞- f8w]aBNE+sNpp#Ղ*$? hɾYS j:ÂS.ۨv!dvIb`xU%Yi#5.PL0bmd zp gv7 -OՃC oˋn3'=4Dh *M`:C?a$X<'$ bļU_ˤA1] )X4$t1ڱ.5!m⎕;ǽU ؊&Y$zD $zsVWQ郷NG'Cm8JJp|B.Ke.;5 \M1c"h^4JzhVQ-ҁHb}O+A3քľ4pv- a4L +r6ǀaP[_ǦtJ">@Đ%6r~&7x;|I߻}{YJOqX.|iY sx)ʹCD#r?V$FR[ƅ!` V,jXK+|Kx3(;`.19v38K*`ɡap,lΤl1e̓Tb~jDb2gfhƣK^`* */ʚs>ϕ"?ŖM ,#:%ͥf#(84fn\1_NzIG| $8D)kZ ~=el me3H!F![8y}{ w TN|pyb+ݖ&8"Yynl"X#Il"M'чRW8TAk\*&8+Q"kX޾w#Mox{=5&ҙjcɺݭ 3ҤW ¯ 4ֹY[۽CnjR Y/4~6=L)W%`Xpo%qIՑ7rl%1R 2m g(ńS6L!zCsfᱦ뻻:\ lat JpEpb$3? Y9&~yN8{iK˻N'B`x̝ 3ze$cѦO]1zM,w#̷Ҿ }ۛqڟUP?|ߔEtIYD[Ezx1 5:Y?6>m뚬 8*j(𛉺o"ry^H3z{<8oxaT]A,Ww).o;SNU'7Q~{}(p {}$og[e.띖,^)&!كGr@"Bv:ҬQI8Cس8xjYt;GR?P$:㥑y:!YMZ԰[[O$8*=e@HNϊi?g֋RI4n0;~:`'ϴG]{NҺ2e3 +>1TސMSI4md:z4/'ӒҒFMڻfɒAH@j1N@EpsL{XGwnuHIK >F G(}C^;CMwB<ILR%)xl111GQٸȚզ Ӯ?w`Z>gE}IGNNn)8AgpB/ >.~#mp[MD BmpZt`Jm <\7 * Fe$Lwfq 4hU!l]bIȬ ه0ȫRjI)gOMfJd`eRqWY`1:%(NAp\ӪHT;B`ݷ-Tw+ lPg9+b?o?-6z{"45zpwj@焬:hIw`SuI?MRLB@̓F1~תoX\;;wH}ħIYv:vڼB]h'o`C"ɮS_X%IqedIG1Kj2=56!7~CpbleZC6\D#ă< =8L7u_YƼ(&WEu1t'Hɬ맜.4xlbEf1tCRgZsK$8@+8>-VqݧvH]O=x a'wn"uafb׉%W|%VZ f;u?۬mc.9qjV C4WFa3#?(L i3B˧1D5QcgR)ϲ g= I9 ^ t\qaI1`:^-sh lH o<s6&$3"#H^d-g `Sj4ؘ8ϫb]ZuVf/{Sfa$fÎO$E?FChܵqhSݘ=(f8?-l9xT~YO Y͕'qp)B.vޞ}t8S 3"@x5֚)[Zt$e.xބԩ1bHmW~+ :lfy1gG׮ʍ!1{.߳w?]mo6+{/ NP$ule4]N2#;L '$9!ɠ}y;G^yo:i/5ZNū7H_#o#99RޤGǠL*P7)ÿ>M¨SG)a|W{'U ,^ |ܤkoG ?Xӝo3dQu*֋mv.+\]G$ˇyq]j2~j&j~h4*nUk-2CsQK = ~ĎCT4Gχ#K̊P+C^N!HEPhZC|\I$f:5 (/FL?4q>*_ BOdE21th:iGH"Ń'Csz2 aT?N@&hR+?=[N{5Y云D n:/8o&BM,ڕՎgQ%,D/. קhEѷꏭ[X3?Ƀ#hVXt_b.9]\ fB9;7DʬJcX'y}s)^&` hJQǫ2%PN8hEon,5fCq|mr2nzdrwp&  P^C&_OM~|CWuo~7\E%Bd[o/rIV(NCUs%S?C0PW~Rd&y]RCIfg /?>㟪50EO~{蟜$=./ I0}uI A~OO9B_& Odj U=R؆B@\24a%KTgZ;C{q/М.zaJ;ypZ%? Q_'MiP4=|El>DnT5s D4_xIJV$|∂,e&5EѢM& t=Ju.S~!pBfaf3f,n> ffm֤^A]BA>%LMF?ƣ< ,~3ѱ7-{A^hZK~c\ܸB@b$\u(TM}eto[p[F\I-R\P7> S\ɀ J`WUۜe fU_1i䣙 V|bDpp18Z8_++7K: B~]uv}S{LhV.TWmm8V+TE.'ߩ Օm~Q;J[(I?ٹم~e"ҀG&߿~K*j8wo4j.lzYYϓ+KA2!ʎ&Ҏ/?J@bl/*P8*=)?.ZfSfe2G381d& s;W,37Մ$HOog! W?ɳW7 }ݽW@&!Fm>r·^/?rW,ӣt R RKlF1YÜEߔ*Zup]Ò8q%ԉTa9%Wκo}{Mh[UZݸy8b2S^hoSXfhQR# ju"Eks2hG=QgVH-Sm08eV+6])xD ߨI֕ugh&? ;F{cMYJp<94wH)ʭ>/4%;5o3Z@?v"۫1h8iߔbn^kK]'smx{LFZHO!cLIQH{f54rۋ6G63Gev1dz;{o43OQ*) [?q%'wh* č[|8HA\X #SPS-|"ZԺ#bT}vԿI;TZ9k%5 {+ # z-6Ƃ/}ZDE^S6mx'ݶs<p@{9ϔ=Go>h4;! |qw/SDDkzs M?q '~S(ߍ>[ ;D5W|0Flx70MUsȫ "xZ\; 4e`̔FP-:>!JYV:{~XKp%nٮ5ႴkDyOF|Qkr7;_T-Aus]M!%lϵ ኵ7=E]ʃj֙ӂ0y"$=cnR(:]Gq֫FѪ#@;:pتMQRpq1ȭ(M{/$uafpFRHUVSFSP^f47/8g^;rie +N!|_>ܯM|QlZQy l8!pBmuVgTZv5Z)QQA7_l*p cV԰434LE湗X( N1 tNd2OkvӊivHn!֌ \Fr0J(&p% AN*(X3VH3Vpf৴NBbLl:~Um %[d׌6D= "#{Col(M31KQ6*$-x_ӊhP31Bik4cofI gJ|%NrMoABYƄ&l ԰fake{Cm~@kvK|A>Bci= LO{os?C4w%PK5wؐdAЁI>,?y]:%FvV0WS$0 ^xMrl( |zj>+؜>\4cKjF bJM)Y1[` F7z0sBf'Nx0ybX̣JoQN>]!(Wg)|0F hR Qi0 An!9`<"^ QQy5=EBGJ9#/?6i u#_!vcy$ 牉~mFGXU Tzhg?~J됽Q0^N7bwU6\ym4%aЎrfuQ&6J3x n_ӡFx--Cnl45iEwsh6f8Y8r?p}%aw07e$<:"3#4ݛv$nE*c;ߍmF,h8~vphKL$۟C|zkV C%agҫ-@qˌn=.GQhJfrf+1PN8Y 3[A'gvO|I7AvF C$J _o"eNsݧdef_oj`|x"X5eBŢݛjz.hfJ'&) dMd<0C|l}+~"K'[2qʯ/6np\Rhpk?T_ FI ߾>]?x85V )ڤYjpJ3O=%)KB 0RBi/`Q utu6sUˀB~^5tHw.3W &~jQ 1Haĩ@3iKR Ä \h`G4vBN k.r3&QRy)afsPISԧ91HY3YE} FeZ`W3TeK*ST91D* a1 2i:˼5"7.9~!%3_",NBI5!4!GAgJcS 33y0#@p2nB= НʗZ7v{'`Kwmpf{&ѹ"SZE$dVt':WDEI4Mte'mqx'IJRJsk!dQtho*Mn6u5Fsm())m6Am[\l۠[awRU3)3SI DEb 2ρfޤ$5ĉLȶ yӗ_cP?7r-0 B;do(->B(lZ#B: jAAR P0T+*!XÂWJ .P%5n"Le`%$PT>c<""i3UjdTiCxǵϧq'DI ( T,`Tff+9a9αfC r)PJIN3և_hZp˅ȼD6}P[$TG{# Nj"k11[ ,ȡG;vpFt@iAL6 ",e2c,sgV`Df&EP%¤3xu]moH+?&/LL6HfֲO~U$%QedNlQ$jm"NJQz@*@m(! F&J2 0ch0΄ĚXK@I.JjB2mCADqZ$L04*zIp:P S8"L*ʺQ1$9 Xb,0[+k('2:f 8Ph|NDrv@3Egڤaqy#7=UcPcT5<:Da!YX GJiA!dixkadi:ayh*jy}z-7IDts "1Gݖ㪿j@H-ROݬ-F\[B怂C(pHjȖGz跤Z١V(-DS *~ķ&ʘXQ8A@c(SkE>*eݟ뾷,kE~s \ GBMB> W% F}c"Z"9% D0@BPK6ݟ) ǯfoV0 5WNo43iNKl>iYs23\ 8CN7p尸>$nr@7 %Cw)-HxϞ!,F;5Rռe@ :J3ƷpfGIDa"x\%(&֡' U`1 ZQ=R1* 4H*HXZͬ%'T'"$̲D:QIzt8'B$ᑥ3NVld%@DHc@1#Z-JhdE7R:Uflk M dʫ}j;rJ޿?ԽyhWQ2+Crؠpj8ܠŖ\ 83,fbzK̺V!+<#bВX˷-2wFbܺh#M_i;dѴp^+t9Z@JE8&ThDtOXq"u+2#DTF8/hڑʒ \fZ+ pb&CDWPH(l,Q1#?šHez;FL[;PX @ֵ i A ̜pwvh^lVKƈ ( JȣWFP)hS/Eph!ZP6"1jakKIGGCIyH3LF-[ nh_>~7#lw=;R Pn-!H5"Gԡ#ړfD(ް`l(-&GQ4,V { \޳%*}T6-mz ̡">2I{ Jn9Io pЦ+v^=vuߞH+С(0y 1  p3\KHno0M-Ax,Ba?e9+)ճ(;K'v:Q9ʔR-cs|>@Z6ʌ(S#J=Jge_X??d?qUË";<'iŧ y,Yq&nF7ϻUW M Eqɻe=yEE`f&L8"!HK0 ۆ6jNJ }n K4``6+{^2 ۘ-}_8TzL{:6ʈsiX3Ȍ/$x|=.; YM;?砣ՕO^ڼq,- _8eFG6==4L0F*0ko!,g0[HrZi?yM#*m{ $yثJ)Dom`{^RPv{%}sI] R|x<865}#F)BA 1q~oyv s8Th8s~ 0ߕ?i~)Qcy}8NvՖjYV_?_O_MaZY6N@OL^N#?yU|NnրNowsX,O+ՅxxWˋjNWu S+4s_ͼ^ŒJ5󿑑 ϊ2zH+rZ=8bl鏅 E'0Eŵ,-beE蕔i!Z6E+#k;|<No_NXsԼ&` s_|\DOaLe]'[IǼ\;iK÷ wihN9ea'K*$9zK*A)a1.Ճ1 tA r>HGEb:q/_Q` / ݶvbm\~]st|t<_]th3U<nQVHٱ=,x~d|rC4YnZ,)UB6Ȅw쉃sX'{L0V8L8w+-L{xr d,5Iҭoas'mo$>w^R@ [xzB;gA%Ӆwye0xN@* xuh:&HSkʪC/I":Rnk F}"|,ndf֞n%ȶNП5.ӝ2pa*K!ޏ .Y>/f; foo)^U:ھ|M2[y߅(5]L/0$Yz4(dOZbJvih+,Wu'` /7?PzE;eHЬ^~EN!_x\$2a,FF$SHXV6R0^-ޡeB6?w0<"a/޽)T!-AEy˼OX& 4&{yĐ;Y=[' #8:p.<4RBw]q^ʆ8g^j㫚ꟻɣz-X]#{2Dd;Oaz[C$?Fii;!楇臺%̵g\d8:NWzNx495wIZWV)ɏj>wI{sf|!Y͈1)kE4a.渁THz8:ŢqOST.5=G-XYE*d ?[ mJnIBj)5w<*L9xmc4Z 'i;n ï&( #LJcl4:Qj-RZK[$ )O6Oz >ٿg쁮Q4۟tr7r^H'bM]#!PHGDc|ޖS9.Z+J w&I˷!rϑeôIRL8ЁAתq1O #ׯFK:;Eܑ:.JK/t߱iw`.:Ԟ1:va\3tTӆu錮EDRE4;0㌠?AR ^bXJl]'F$pա]Q< =gy_%o˟S&z7o :,,-uJ 20VыׯEM,܋{jL,"l  *~@2$y q9+|\Gˈ~X.G #MF`A&,f$r a-좄taU5{;pb_H1+Bb@Y ,б¾C8Vsms(DhqQ .;?+ vlƓr Kӹ0wmI9@fA`wQ.nZT?묿\B _Jʑ xOa[+jIsھ/USGZ}}vbsۣrPiJkbn9!X7(ar^es i8|mY0 Za:qe;15N{*K W<;tVlx2jk\foAԎwݶvwL,˶M,i ƘKEX,V1 ;6-],V1xЛ=%8(=;MĔz^G{a{/ݐyT>|Q ŠYFˋKN|=At*KĥÈ;pje6_,&IZ- /o_A qʯ?9)槮s%< fn *ub> S%d+O23/uW%ُ?޿_~W|s7SVsdֈs, "'Zp?:~mvX[?m|%Z% uw߬_ oQ.@ ]?Fx ejN(s%_b=͡]7{~)wco*U3 w[;./[HioS>.Lj()g&oSxg?GaoBYчl F1Y K|iMH`0$g.(N_3K=r-(u)^ׯS,|8ljD§90/Sų|9–zyX.7B##[-|,+Bc xlvtlXens\rm]\=,#8s)b\њSۖnYb͑ND_twz2M_ĜIK=)Xm.y4?-wv~9~~BSdϽݼ-n j*Ԣk9>YHe:jŇ*2M}e5r·g@L:f.{2xM,JZ)]z7Q͋cD,{f']J U7wFq9g=ՙh6\Q̺>[ޮ..%ea?|5ZhVvhp`Uӆc^|3շǻP4&ȅoS-t?ǡt:ĕb0UZ^ 3#՚!nC:pNfq:G/Z{xcRI[1ÈiyN%Sެ~,ަ1NӣōJѕb K\;.X-lCghl95'}҄ ,91WobRr&P\.WI+RoRvEo+%GZxt"3,CQ$T>]Ҥ\!/1W> ~B%iM/Ct,|}:‹=b4vvHd_r~%횎1^^rH_-1ƋDR储'iͬr|lvmcݞjGBmm!Bm ޟ Bq\bJ@̀^1 q4]10{|?TR}>O6+^f]i2LG'7+ ̭cMտ{6k Bp@5+ݠ%0M 9ܵ+j1Wi~!.@V@:~nf*/cPTvcbšHN!5;T?(\̥_j ثvMɯfGCu6΅axXF<0Ҫh~5{DIl%N"%n`e=Jk*YCqK$2{ i c=Sp4H>?UtɁzp_GZ+nBar,  ux R)<ŏʄx.ʼS#gJeXQԅ3j"dFާ$@3C8s(1(<: q$S\)149gB1ɜ69/ 2N+NyA4i?u흟(C\c a)5<ӔA=ISi4 |(Iib3fV.MxɰIC+3WY+R5A E>Y(wuul23:Ԅm-x^,7͟f[%(42(471p0L䨁 5sPru5Og9 j$lY]+?O5ƨ9G\gu) e <`8bȐ9)Nw))qu䚐1v [&|l?Q(G#e;%4#8 ːA 2p{I@4sR]gk 8XdrУ°3b#ʡ 瞤)4S<6ea1p.)0e{EA4D iP4u$jG}\\J3MϑT#Y4s 2(P{h< |(H~ǥdBt3=^6Wy>&#|1o?}Sx%SCPdS(XB E ڑ\SHJW&~Q:B oSd<9N&y&ɓ둜r6Iy$Tf]H/P>ܢ1j鑀)=|D]oo3Cbdslc S#ڝ]]LkI> `G!gqo5h.[ê\cB:7fo 2 xGbv50wpXӍK@BsK~11΋ܐ&f d#,j"b6>tlXQu|, :(A#FNp 2 Pfs?pJHgdn*fP`ﷱE:ea4()5>\LN7K}ʬIQx:^Y)3;nc-QNIΎ1,eK{k t`8C grq8тܝ>b<.l/,2H,2_wyTXlW,ܥ*4i/2rz31Fcc>tVшC%t)f}tET;1 F$2)3kX{ʔ'MJ[u,]cn4">-cE3JM̖cOGS\/ˏSDY*PH*:;`ƨ%HIHr5b!%z%BpKvӽjBj%iymc .JI_P?9xEizES\/ ]CF';Ƙ#Z@M'ń꜁jYhgڜGTcԚH&>|l\mo$⟣n>zĻbآPt5r1%KlNw%6Oo䤲p)^S.Pp`p?j3%YP0iaA0;'qŒbN Nm>?f Ę$qc邜3`>ujڌI4cܨ&9;Nto\+0k;U="X֗  8H= 8ޕ,!z嗇]m32$.S4(ڨZOOHr:~O2wյ9IIM7ւkzFżwg:DCMYSi&#(uz5#%jPAcW`pI )5/qݛ_dITc"뫖J/\)H 5˟~|#<1Ƅ=(vܗIPM"\=&9\;Ƙە%>T)cLD:0> uϰO2Y\/dtƎ1&H6KMXљ;N9 |)h'aJɎ1[ZH>bvޮarpIh'FʽluGh|,*ioVwA}k5e]Xe?'i#=ED=&=ofJ_/Y+2),QP&' g:ƨMm/ | 5VS밑utؘvوۋd>N6t+9G^/;~$)_+8ReGoŔҿcL O疗0PZGLZ8$ΐL̄E8塦6>j:AfILt@c.;)ë]SbqKzG$ gʚIn_a̓7PdIPm),;nNI.ixunJ&jg6 (H |X} 4&} {NA;\ d(܆a cZ.Zx;,\Sj(:S_`)#W*V(XT12V?VAKp`3Z~pioQw?b 9rAVYEHGFylD9l"$` 1լ', 1Wn85xdZ; άQqx(Ŕ2% B|5E|, ?~] ECwgq9eq+(͇'C[P Q2 n-X 3 pQƱ&1P$F YZx<o_,tL@Gg RILL gYRRT:k\y,ي@G{ =̈́he&gM 9Sk[ <&%N?v_`4ʔ(jcA¼DQuTR tTT|wm#(oQrt7텏@%@@-< LǂSKA=+m%6Ќ2: @ \N5|X٧Ѣ9Q%ǑZ C@Xy#hcGg|ա 9(6oеW=kOp,{qOFd;u.$(:oFq4 *-j㰭PiV('qq9ւM059 ?).Е`1 ~ u19$^p 3>zP[+úAoxQ`H~OiVb?֬R vFXQ ̌(Z*vNy\eѦGZ~tE.VDYzQ!WO['ыHHeh4FR-?W=^Jbs5>DpCJ-\Y(=EU`QeUH䈕ik\ʓSQ0ksʋB>}8ڤ G2" 9^ hU-?W=c\̗UޗhR-8BhQ\m+(MQZ y'z %m-< .kcˊ1hq{\0KRԭUedN(1'$=3 yz\1XVܵZe#S H)54[; A\L\hڧ@@n&@3'7PGɩ z\qD$)s!bd$3k3URl,/NGN"(ǭ Y@H]'T)`XAq=Tz־BygcEb;TkTg4^5' ۚos͚o3E+HG£TL|eCZx uI"ҹϓDq0FCplC; dBjKiQrjDȷ\.W^h-}#vx0J4]a_4IOTϟѯI,σ??t9FvGLUgA5Fa]`bGԡ<(F-?W= ܷCCqbz:l#FU.R,c]f6&8rD ǃ}`R VqxWDXHEp6Ǡt%OZx=Qk~KgGϭ7${}8i3m[qIYӿs(VxK{켯PL0셍 sACk\%EQiϠ-< MN/O&]L4xn }Qt/-قʢuz)D(8kF (3ESt-?W+!G3[x<Q+Z 07Pg 8ĜSjjx=ۙ g3@ϔ/a"CQQ0hC# "M^:QZx %·e`J[x\Z-(+bt@R6}hi?NJ UR@.yڪJcBPV皢-?W,_.}1 kQǗ 3ϬMt'^8zhmzSFnQ0v=qh>Fh;C]laR)m\7es:vF)Bsio H櫻mP=HNY8M."}T)>-!6DB<*hB2#(dN=g4y0j}-ne <ja[;yskMBm5 Z [5.f #I8jgV.ūۏ^]|jַy5&f\* g i v n_u:hL&i>ͦ||-ь\I, HFi`+U) xv67!ie)S2zWh8V܊VW[+tlIQ#tkh7RFr! `G%6l.~iXt>}-]{#{n2=[\XoZ#$w*2^ʵ? A[0P.N!95d+xiIy_ٚh@^|81AX  iU0r%^f:6߉fi8W.Зw؞>1WO} ž[vrzrS,d2nm‘wF~MƉRn.o=b KLY9Z5l,mhovR"2=]Z W2 qh6@XL|<qՍA`pAb#QN*DUWYSA*/s^M7zk ᷵pk\FO7zdQupA^߆I3ƪ*aA^ v$͞{34e}ǸQE)B~aS CT1lKP%fPEuSk`x n# nCnwY}s?OWX=T}l)љ޿ .˕L#p<`vߴVb[Q%Ocz9jL+ +i("'y/cq+%XesX1v*kg l8]9\&[v4btb4}3}k7;dFGi-0}=%_!.@ #0RM1̿bh4ygg8d.{:ٰ&}VǸܰT+oVb{ӚTϾƎ_Pro$Z/eu*I7Gݿ-Y74_⣱o*(:K K 8_U$7UQJPeP)r PrR *k ܦ>+E{'~3 P-"{3R״| nq- g<7![ΣnZ?,5, ߟѾeÒN֒q%Ѐ+'z#(h2O~OxP˦XOrsvyh @Z`|Hg{J7P]aAW^rQ]W8ӵX&(1*yL0AT &/Ig66իpoimѳp>&fxDtsxi'|WK8,wu:NXQt?p>wyp [ lc/ =r'/P[+&+.~ /p0Jw* \ZQ,I&BQ0=۵ѳpwR}\){Abј53[?$SI0?_q?~s /Qp$d:dl\)bLju*_[P\5Jc!I:! )F!BiLKpEWFW7{(d1T׸FQ Pp.I&aqZYgutYyh kcڍX:>vy:q,?ɴ@}Qvb>⍗@)[ߚ0*XBEp יGGB O'#x=XJ7XJ6}]Rbp>%Jdb2Q!TS^&Sfn8_Ԩ)lVCoV_V!Vm*<(dv+7m3BσW&᣽{g#sv{%q))}Ҧ1iK G0&3Jt; J-.\dtC@z'%$F9 JESAdoIB8c$V<ߪ ƭVUURmHqÄw77{;ɶ5%Զ,me%%aKd5,ֻ@P5G'98y@nig>|) `BB|`rR񂌲(&Jpj(7y0TO  w8x3u &|r a񪊲2V?/x{d}s'AF |bBW> dDO4JS}E)7s׸'-QD!%L{$5iԒ̥٘RL{+6d!|:c6>2Zq2`0EBoLjD+e:t~[0{8}ok2Zd Ĩ=[H+-4BZ+8I1=S 2>W)00\A KTfxF$sƟsn1!G}[;-%K*nxj2*m)=rz-K hP Mhy"ew鬞bG{9!44Cg]@u¡[yglM#W}Y0˖ yO)! 7$cK(z<?t]U)0&q8H7)eA)<0@9!`܁B%raZgoi Hb>/Yd,C@k -ĽQ "B6Az}3toCE0ڡQ@WL2X$m=.2ne'"Pr"H|cM%U`eqbU*R0յ[N;֙{=E('L\/Yqwen0:c'kW% Cq](.).$n1:P>uӀi0u(!HLsƂDi.NIL:{8E(JU⦧#Eqi$B BG$(m@s,WJqGYiL*e"E&VV; }y:>^`Ќ{R_mOU!, $9wh1,ۅ+SX?30d쒈TH1~A*?vNfWdZ]Su#%%cDUwfs=_BP%d# *HoWx:{q4 QXoΖ0,,kS@f_[x͵z78#!Ph(ȑ?7+_^G Ei.)٨/GdjKn4?Խ7B} 5 AcH8>շ*3$B?Vo~muڇ&XQ_ͧݫE.ݛ]VsmuqpǸf_e/dN{5(z( fru`} (:*R(vs +1RG8aunj崔yң:wD&Xˌ̐ ч!7vkzנ6ldfS{pŝ]v(A?UF1SDCиB!-!wAWdAN9*f(2_o&餸ԶWx B+k6/DQ@ūţG.gO]tɵ["DG!_89\ͩSTF&5;" I3m2潃~?_N:{5q4J8m8QqqVDX~*;!^x ^otpj=Fjά EW-I=9Y,{K܎qs ʉ0B~#ĺӱe0kNlO("Y` [:vq4Jp(@Lrwq:; G EiFIAqugd7nM!#nB^{\Mmd:P]ģh5"jWܢDy5Q*Z3g +FJl5uBQnR9S̽L0)3Vƥmcv9FQq8S 9Skxg4Ly8{zOWpX7D@xFj,+˳ʲ-R,k$XּM0E5s^#Wxh%d~\ezoC`F#p]P`GPq#$e !+ (fSϝG'ǑcufE<ީPvm%=M^s{F߅rxCexUӜxшT/3U7[ĜẦԤ˻\,̷$*h9LWc5&Қ pjb<Zۮnyh>(~sgt HeQ\4V\!n41"Idm {o2X EgM ]3Uha*[#cz>7^Yr?CBrol}=ݜ#R19=0wT*ǰvL#"r \TX祲=Cl(K3Iǟ:M !?.əxsg|#K,IF1(ki~%s( a*k|" igq @׷퉗443]2inD) m;Ncڹwp.Bl@zC x@h:{q4J?T.*(lr m?=FzggC&Ct *;i>ALUɌP65u8E8gV7 BlsbybkudD9m9ls>ФrE];r5ZtY0ClAΐY`8N8졲Ofru F66 H]cG j \iuŕ@^Q솻Q, f,!NdNaqJ=Exxoj80ĽcD,a{sxU%7@WE((ch5;g"P©ɥٺR)즮Ԍ>܌2_q4JPTPT K7|ڍFdH=yvG0ѱkbMzG]*oZc 2`7cKpQ~w X~顳㍣QB~amu,"hR!uW R BoSUdܫ'l /Ag/8FTJK%w?vg;.WpSO(m2VebrPCgGKԄsw'pX#,OM%>fkߠ}>*EHEтWO{ (`Xks)Y̥)˸ VR?]V,k0E{7O0=e̜N@5[ՙ%u(BQ\d?$Re8 ޸Y\SOfr¶M~ =>WLSy-Sb@2jkP3vOqӐz,jWmaMy{|w'g}a^? eY70{O?L=R(nL3e(/+qTu|;`M~7'RT`YP߯U?t5C ݀f8 v0G_hjWoOna(a(y/>]v /J7Qq;nƒ'uu;[!eq3w\g#KyYNsץț16nI:fođ ӈk@>ܽ t`̲_][*e Z5!5k?L,i [k|FKoDT|iyd 0o&> 'q0=LfU<2vnQr/v;lV.Ty&ìj#\Nquo8]|l2:I6>Me?z Av7dNlylxow_vWNڇ?86/֐R@gހm+m$GPO;@c_z0`i`fV[Ju,oPQfJI*(tɸq?Z…WqpNy6߿U^6?M\3^QpaZZ#bR%ȀfdbjφfoY)Vc%Ave*Jl Srk_qvDK>K(8Ӡ/sxzrY@G4E<l{ELƪK(>w1/y PLzK}y7*k:lE/ZQB(ig:ʗX?C7NreAY&7V͋[_3w}/^S\sx +aoPgtze8lr~"}Q粯l5$ai<皇w_p%شJ',=E/ O>p^?O?V~w<~dϰQ*Γ%k4`p3NE[Gb&,3v->;8@{- g 4ܵ܄[QX?ݼ׸D B߬"`r|gN6_Go>{4 ˖ j| EqHd" *ͩ'5N774.{nem&mh:)F9`k¶;~qY>`a)찔<\֯q glXHE$Jixzu|=if{)yN{YQ*[% "as-:"6'J`PN goK>8ǼDemՒ& }y+'IpB抦k7 昔iܘ }:Zgma 'M&Qlj/쏜 K2!_r[V_2.!WU̔$(_5LSyHOy]rP2@q똾EnD ؿ zlMj3$Zky@߈#@,fv>og< V~*Tyjo󚁲jYFP([98BWpt~1_h G5[Zb|12:Z/fwSV}{Gn6o_"V$->E:oSesI*G薮O 2}sSd[VpY|GTo2R 6U j6bTZ#S|#YCT g3P.k[4ɳ_6yx{ATJĽ!eT+wgF \~yFX+?>_?m|{jj@fY7q _ٗTE /`c# qA9r;$ @$ ltp>}8:@ 4Ro 6<*c 8 ⥨hLdX<Ce\hǦI@Jo% 9ܯaq^P2L *`OcT1(bBu찠kj҂{mm"ӯG]p6p-(p]YXS郬*% qIE‚m YgA4M:_ܞZ7$JuW7ER}JQ!%>P7T 6,~SW}s ^Qid}$DL ]=b+cL.Nr4lrw?_ HzQ sIyUaH% R~G#'DDe˰1)(Mx|gvy`Nf(Ӱ&_sfxpfLE aDtX`1 12V_S,pUZj\oqNȩ] qRON,J~mϛMxr]v1#a}55<ԚZ} bSE mxZ(Sl4XbH fmzD+#1U@!~g7.u<ykWc%\;U)>(8!IdÇ6mzlrXۡ~$6I *E9£28#>i-iX]c& Dsw!C Ah"!a"5$ǛSS>£28܊-4ES;OI{no)#<*c.rWhi+<~R, ZQr[ʢ̯VHsx_IB'GxT> XE&e?QltuQlxTgy,Jcjޣ^Fj^7n| \Fl*4Ygq &CD4FY>q!qJcR ShuR#J͕9H,mF/s(*]i!gmHWBb7 w${pH-pA/˔GY_5EɔDM[tgрMvG& ?r5"7#2qDD?8t* [\!duvr|:4 ߁'2 M[5o rxdrD:*?ť8"| ؏mAXBKc@b$$/DF>fA)wVȓ] ]G>H#.1% */rp1%LpO\#2?!<U1Ӂvsi͟U?ؤ·S2p#i4DQ*5W `eƅtx)MTa]p*-e"3\4Uq64vKWcJarWd0#TXy€0XKk) 0],fLKnL D#R`-G,^6vĂIؗ#A&pD&=8c"L uF6A((pȭ^TU bMc o쭅0ɇ`Gzl~?V# 3T*8lÚ>Ed` RN?ȤYp51cMkwŹUfg0prLp+w w)oG#Rb|: K`SIAO  KfysZRFDQs|UM2#6}C^Mf,>+)f_3 (!s$c'4]ȉ l4W2Gdk֬w?M<sD^{.d`22;ݼGdT^Hw+AddYxd6/:j?ބ.W~ۋS"h&G=#%i^.ƴ<:#.qhnxZ(=d}WQ3\dv6 3z] f*G%N~,5xcO_tQi1>-F8j QFmL[{Uo5KxXQsWĉPmHѧ eF2FHлAGd!pwvL,~lmק#^JYYJ%^NAY GndTcr)<*vG& f1R\&Q YwJ:32=LGW}1z:`)KM@b1Җ!$x{CrU+ٗ c6`#S!@Uw۠ ʂ Kp}-ؑicDnM-Jz~ %%t7 .rrcϋCW@hwakڽJΞig MviZaG r$CiHxXqr4$er%$E2Ǡ|/!}hu{WOk}cn} Yl!01 y,5-Xʥ~,*\ iDy7x^qS=M`)kzĵ 5IƔBsWKDSZE4^=%sj>sUWWY>`{o(-hNr0vk;ߤd)"nT㫎ռ@:6z}EoenMA,GlJe,j`UWz]NsJ X3++>PK,xMǬ^Gjg9TU^_4֗r@p1linӲ~[yChNc>|_'ޝ";z&U]J̲ DtJyITe9N<7Y* 1Bd&ɍR f1?~^ Hvq=-egX,B7-S|\fO)ngfR뉲5?0K[Nj gYs6l9_nKw_>W>E-5\);\g]\KF(ߴUKI=tm]\m^,e|_lܴWrn9aJu}ҴZ`)6 \Ζ+E5~^VRm+S?PGUAnΈxl3/4K;[o pZ>[/%+A:(+_`Ωa~"[g zt~[Mj}3ټIIiu~*xY<`=׺sv'ɀHoםp =Jz@oA~zOS?@.lKG^~;@wLp{jWnĝ8-HAyNQn8'qΘIJH 1BR$ɀbq{Nʻ~H-`cUxGw/޷^/,Iߊh{ȑ&9_(<.'U7}âO+YǠ@ +AE?qJ=s>oˮod|C(E}i&2jRN3缿9"B`nݵvybsN'C7W&İ&hzOGcK]V̽k.;In^=Gnm dl}sF? )XѪ-[NrldHI0Oy8O: WPY(C'͍i}wێJZԨ-5py\cR%ǙhѼTln!\[L;|I=P̊N'7P$@l+rZ֚m ޯZ9ވ_=-ӥu>em%vs:ݭ;ݮtCFbkrd`;?Io5Y;f?"?W3~w3~w3~w3~Τj;O,( em0(L"2msSr"i.,:Qbon~m:]pF1-c/Et1Y M)<մ|eu$tvl^io7}OM+Թb9rQo>O:HaY{+IFOq[s n9[>WOiUtrl u3|D)L $:*TyVh[:`!<.$`Z (" &0 [,t{-!| )I5X vB,';\482JP^d). )6 ŔXiB'I;IIbJQKXx"=vۨ'yNM8|8g")-=GI9Ýa2))JPT, fZ:8X]vEbz DH<:\ױba7*Kɲ"%W ^K'E.޵6r+"ee, x(#KIs=bwKjRjےQSlXX2>!8o!p 9ĿX:τB)bc$qYmcLD"Fp4X vx]o1olڷ+jv0dseQ;6@7)N+b^.y ɹ_~PMRd VA8 L.&EdT 9Vi~9ֹtKך a 2DIELR0 L-pKt izliOٙQ!>N@>',.``ȱA4@ h &"Z &HBpN`c]R 5*ت7{KQBu#>]u6^WwJ?z3|vc8Pk<҄{?Zl]yfrvGB\-SE dֈs,"'#tpV|ֲO9]9x(79K[Z5!s&bbIFq ygLXO 洨4f#\c%N&*u9u9u9uԱlyjm`BK%E}0$`9:y1c#ݳa0[ $~ #< )\2l )ρ9O mI$x:K A(cz[zIϵLD&O'O9 K-[~c=' @=_'nÍ"h|6 q{/y瓩xTro$Z3!gE8O'=v yxͿ}p= g8ԈGːc"I0U[4" y kg1s@d^v@d;&+'O0Zm*ַ4~lT￉, ڻiKr6VW3).ө7y/z8lm!p?bͦFo][db؍g|`lȐ#ncBZ$J ՙg"(IȣMYܥ:L{N'#څh 0Ne"Hj9VHtcm%ުoH )ūSS.&zB}u廲M =&3~"}vhrFƭFxI.:EKTx,Arp{̙3|6[/#V[7DžEʃ֞)-vKP3LߝzS$яBndW@;-9h^O a]j]sohAT-#zrrYA2$sA0N0w,AB[ ,{k`(xy[1tWՇɪРRcwx0Ie̿a;]n>tMr4xGRޒ_xTA \6G|}ψ*u+Mnp3]1ި<Oȍy-Zfoh}tmB7z+M"ljE+u\μ/T)<>ɀxWOhrf(j18<87]?=a+Z7sIz`};xnf:_W?Ȇ2N fj<]|!>J3rỵfC"ռHZ|?#$h?Ai;+eC^{74ymQD5Ncp0uKh F%R 69k5I-t/>7%`zmX0#5E ~$L09N־>nUcmMEfm}l`eor7vye_Tn??`^?ק*ZXx#d+zP!GAĭA,!l^d(Apڻ>[~yZxBh``$b|+CZ\zH{ST|߁;mg;Er:ۊXs5UÕ6 uq)ߌ< Fa r,fm ͦf2ϮAL/}8q9_ 3R"%5T 'l6 "jO] LPxP"p@F Ԋh(/^wѿ%q#Br9-7T 7y,]ޙ{o=廚šUv=ɼ1'2 #ƘC 6O9Wk%k E\JowWI5Ѱ:?eߖ#x٨ԶioFGZ&4sf.;?ËW߼(x5tQnVn׉YoMmnn/XUӕ5JZݗysqQt.{ޒe˔,D p/jW(ـi5װbqZ}T_% ?TKWtݥQE.*~9 btϮc'")QEF_]7J/ИD XkY|Y2Q z]ݬlj_;W@g,?XYE.z J5&Q,$S>X*hXt+ s:5X*i~h6>d|H Ij؂(G1xNY}dq>ٕs=b7ViQuvIJlDPp!&*լnowC6^޹蝋L:0LuŭF;3g;\ԉn͸BGS(nIOXȨWF|ؠ)W[y̙GQEBkEog *)W˴CS!Z#@-r ЏQ1B4eV{1V3-1ae)XJH#"^FGSQ;~>4Vp)FOXyt;)Kb :6u$E`ue E[t/-iz (>(#1h$MV>h 폭TԡNxHmMҿ"K[ri&'5"3K* aH?{WH/ ,eG0ǝ}`v,$U RrH)-.n;b/q-&FF&6%NEEmh]Rf;A_⒵F'h1ހe"qcQ ,c.F(hH)~D`gD/J@KePP!1O hAs9̩lUD E_Wnh9fM^!AfK`w"\H@l{vKPtAs9QG`q3B*@R0l[4csCXE(E ,}NS+:pJCq+Y%w[ɏn,'X!ǨE` F.$[g5[KtSrI5NHɓxHx$>y'C0~r2]RVR,n>R \]đ$㑔:B,0 @DepĆ)0/,=KeQ$(O?7;nz}9_Ӻa!24&},W2;y$T(:D.x@74jT|`,3Ùpɏ#&F1sၥ`FTw?C q:BjhMrlK pΕt~ns%Z(X]e+a%wL*3dv11/"B3Eߑh ZkiD)Y xg)Y xg)Y xg)Y xH^u (*SEJ_~̋韍w[A48|Gy~_з܉㰊0S; nǒۼeqU)q*(~GqgsO]aV a9<_f &2%6OnNSZgc&>#z>fT~]jvF#fRNv۱NuO#2[7nIEҚ `1$1TK3&H e0"Nrin }#~myUk;Zm(bpHK}|z54=f3zϞR@,јSsy?Rdz T[P^Tgnm t[0_QAA`{T9.y0RV K;3 .62&gƀ.W<2xe)<Ş8b>QGT%Gpo=DmaS@;b4-Va&(B23@z&IF lĝpEq<>ۜ {8b/sNn (j1Zb#G4-"H03~*~ bXN 앐pɥhǫH4Lh*"6FB)+ gNXF"FjwAt OcƒϷ@ۋɛt~ak=7dpUd<^K kנJp2h)^[tLّJE rLO^hgnrE}j:j"H(tvA:h Kך a 2DIELRYo0ZM-^D +%i:4=[ >?M8zzOX]M z"w t1j@c(0R0A"s'ߏ?#ٷk[hYxb%~ o3]8u>z Fݫkk ҽt_Pڹ<,gڛ1|$w5i;:#Cc2t9ǬǔY F{S\c澚XZ:qϺ7j//)3!j{|,i>} dX>1:.c%%gpn# a* =EOb@=R|]1 b ֞_o[|DĤp !gbR RXjvD9"x d9@c!߂uym)z?zEy}P,hQh=j ۦl"EkrF\+]4mFY^XwRz-FM33/M٘*Fj&&8a.x!GB!`)3kQٛC㰉.%zz{Qq F:"_Q$wAo!ZꉏK+ӚDpJaV5v m򽪄Rghkj ovvߺoYgO rRp*9mV)‚ [k 8 g)qbD2KY(ʡHWGZ(m.ZVU˦ߧvn'dtVyzf 3|be4 L*/<(jJ6rDŽ}ka@E `*(SzSKt-!\.$dhdRd1 RqQpFL8MY\p^($E=YpO}XӲHZ"iE XS>BTmQ:s\+wDo9ٿɾR|_CV:|/ %w޹ M}""R_!חߺѸLhσKͅt!=l%5q6~FE*޼ϳIn>9(PM!cFE JF]Nfހ,NoOh8ZC%f'0M? }{a\ҢM3|tNzX[-J߶ͬ]ﭷފF{FieEn\ܜͶ2ۋ |%+ou 8?tk m-c0u~y<\VH,YrfY m V3 śc/*P;YԻTj߰ ܘImL7m]./;6_]-/8#)oIPWa@/@. V#3 Zm{tm43=[vdo2C.2 &^{ ;|,.ok7m#xl^/40ٝt@aUWf)Դ뒉vL4}lLpyOf4 =6X/զ͉4ak:eK-ߟ5ڪBS2"wCE\V(@`,InMĭSE {#eC*^x:[{᰼wnu_8%f+JE5f j5gZ wo9tzwF}O0 ϧ w -h~w'M?ZqƱawwc䷛8m]ܸs7?.->2U}TKsE< \'Yyƪ]5_ܺ1vEף0<:aZXr+ F85Bn/^ ]0.R] i8O y߃ /R6'=6 gP"WKN[Mywx+cJ>K6|Ku%WV()5:̙CNՃ)1ZS!Զ#kd!^] Y ji4}Y)~buZi0X6ʽ?F_Fo<ۖq u`1vt< G???e29Z%K70UtwAw!'5dPo~q2 q?K=@U{(f@UzpU=gR}yAf8TւeseB}CkZs9秋yP|{A6gs vYPKu 5F!- C%ZJtq`е]/tմr>gX=_@:Xj:6yᨍ 0:bJ/N.l9$?}p(05w yt{:=Ql=N;[B7zf!% 3Bs4sm]ζz4Ŭ cYNz{JIm<5®'?}R cex2D> Vl)8Sp wݺRtnD(.5Ԇ2xgPPJvsU/3moVuxx۫Wa?76=#@X(-yF:H*0.ۯ{(m?E3E<+iu?.JB^f`]p=^)N3 {XәdAopAN]β/EvA)ø9++-0L3 RlYεancJX:nK*y0T :1/9:g#wD[_CbV)db>ؼCtyomؐz+ ĽQđj7砊hF99<(z]ZV]iHXG"\JlB Oc*풲KRk hz yֲ+yŒƏ6&H=Tk obKfjw+"7,Rex:/5!8Xbzy[zkz֬x>$9 jLlsy`ہ) iwOn!91pb6Jk)stPG ݀0c҈״M\`BoP2]3SP۹x{oj74\֭]gR< eM ԛ$"70M2U^^ξo%Y}]G~S=c&a)'~W?1_@N@)QͦKj>S[vyvn8Xy '6RD{5T )#:Rʵ(ݐK NHI)uS^A Qlk~vO4 G0`QX /a4z !z,b`35ܱP*jUT@mm5X8@{nK||=)#0:S3?`ϕA`5G+eS>\>CRu@~$eqkQ00p$^"e*u֑X.nwịE 62( F刊`) NPL2-spkkq ͣqS4ݳ. 6E,ey"RxI+%^43~(Gޠڑ S2`K.- ʠ]-Mm.@$rӃ oib[Ѕ%,s3{B/x~VLֳ9G]4}_M8$t뿘Q^_n)b^}Ȑ9GGXzex62<{ Y&m]|?썏+ai}ZχjkL4¥ǰh}|:}fmO{ u$ž;xK7gR*zH뎖\892ַk4Vqfs_zpqxR*_[n-9Ab1"i\ʶof{GId'}1U5ey#n|T:, f屧% X.oP;y9%o6 9dBj6:볨+B:z?;yz#*O}ӇJo6d}d*qk!^PB䈪K=Ԃ.Qhguo%5^<.p17H)A[x^Mzpt|y`ȒdO='ʗ!D!e!x0k5f,`XާhDk45[!-og>˅^6]*fb{1MI(e@2Q:4舙fnt-,Qoӿ< @8UaaR"Z6[n W|DRW8e ݝ<-ޏ >ߍME˷ݫ 49ޫqʝN\SarUܞsB(֢t](#Bp).}Ԗx9m ?\1OnR }suW&ɯ ~I'K zQk %I.P!`V0/2GJG:i]IՆ X1;9ia> 6(S;$0 rt-0!<0GL1mpؽ :{A<FHH5pFBb%#11)DXҠ8KFbR6xlsiuvLN{v\]gӪz;;/޻}]i-ewz1rXf\MYapuY]ok=+ypR@a .(uiJ4Zο~Zٸo&~(z_z7,*3veF = g.)L1 lL{js`1O%Tpk^0S )`0Pt`У0ιe1EFs%e+Oj"yN.k˅)+0޴Bw~eQ1OEf-qar3Td~jwLsL)솚L].nx3u5] BoUGI9BaYbKa0~0_U*&Cj#6u@u2#c~{^ċZ0W+Y40Rs]?EBacxfYN[Ln:<~u`F˼,).vrKcGOsoH{I4 *`)*ndO|W'Y:"%LLE8 ~7Spk1W#t#? zǯG$9֎,4/mX:% xt^TR7)djIk-b$l ^+ÃU-᥃%e]< ~!=Rj& fō):w\N57IAq5H^(CVZx&-rL`&לm?.xvO0QTl;#%$KTQRI[K^Y{+rVRDkU Y 7ň1?AB^K?VoK=#~-WyЁ%kyO'$Ȅ7EBAwgxJ AHoъLEk *RF {oZ6MY4ZXbEF = /z1r<ߴFӳF|#f2pOFd$Vd(1s^hO|pr'x#_UT^\cV/ޯS-@:[K#Jm=j[>(aتÍ3}!T>2ǻl:T0y[¼G'JhQFֺH:WdoC< i p&(?.xA;&,`*b*Zp>XM){elc0gTdsu PRΕB 6rXw wD~&bR5ۤHT7EB3WDJXi۠Y1tU?EBjwy@Xj,J =Ki{:rdEvu;d-S )S$tﻟfZ aM#&KA,eu?ABӁ2sa(XBklH$DuS$mDF$1TN?xk۶ٛXVHʋEl7kk}fޛ;h9%vx}W:I煏n_ڍwҭ{0>sbqʈ~|e>Nmro= i?,_X͝ļCkݙ (fy* +0Nl}ۍKpT|Lp.poۯ3}T g9&)8ds=)NAsNZ+7ZX  gKMG(QRbѕ9:#J՘4}LxB;Pm)V W ]V*QL$4IJ u5Td3SFTqb+:A* aLẄA U۵<:䶲!*z.9u_6|vuZ4xoU.\[vBCWOC܇t5Ie|K>@s=.Js ;qO=oK>g'Al4w%>~4kH4-I,x!t$t\Xy*yb6""c!^zxl,R/Byl>EcU)myUf͉Tc4 ApGZvӂUK,Bg@T"sZjW~mAfKr#W^x~oi.[ylDh&_B3=(+Bxzߞf˧w:gvo;9ē O^W.vzIyx\jt'wyn0OIun ϕ%9[aX5Gw xX,a G_O+swVj:C\*T|#(Kc: E٤Ioԟ/5Gqq9[>Kؾ yڲ%Jbm*t~ÓS5CFexvkA*EK ܦ9^Hɋ*u>![H#Ҧxѝ5gοUUZKN2k뷧!ޚV-$+8ߵ뺀Ɯő.:FhirՖ$Mr'eز( |˧W P;lᡋxyv.EX-Y:K9jKLmrNzqsߟ>M_16cTw@TCUPһg~O_v8e\4ue!zJŴz>d``N rj;}B{#駆bTgG+omha#UΥkr;H 7YVQ>w{|Z6ow/3#ƎI|3ak\$.o%(0ΔjrchVKekRkŒ~W~nS_}ʟ?F^<*$KNO/$<l_aZczȑ_) i>؅g;`zfvz RJYren50}uNu4`Yd18?"π& &19QKVR1|%gS 32l\ uxV}M+j62Hu&HE(";u~AZvgLGk16eLt11 `*|B :TBS6҈)**z= QT@.Ɲ;?xsL(QmGQja, -J))$d^6K`% N6efZ}uׇq ĺ/GgTsTsA˼k#i <9|cPd@UNwY| L PT(sVA-m=3{9A N`[l ϲbe>Q7w'lM0U*kWS Dv Ot Q =\ ~ q҆2Aؠw9洨R.ƒ EiCLNJ]q'ubY3oxUЮ|xl}q#l[4 hw'a=3X ;w/O߿ /D(UJ;V,ZYJL%0"~ ʞYuEaR 2!Ӊa tҦ XH>bbhp8%b"XZYh6.nNM znuJ*=L힛@_+̌1Tսhpj Ӑy. J,$zQZ5ʂOJJ# 9=Jy,d o!)ʛEy'=09v'O^Hcpf .%9S|5XBBnF+Or/!s1V)T1buI^N)/URz5Tp-RC" e&y+w;< Aԇцj+)0?В*BXY9#RSFDDHIJ+CәcIX:UNjzs|NۇM4>Z7CCη]txJЧ݀E:,5!@PGf[ƭJG HzXv,3#\07}4тQ"h@QAsglrD`p'm85x-z`:hNfGmSZ4ix!_~==yMZohf{7S)Ttbyˍxröe'/TZpi9IA_E_*>%re()+SL @/_Jۯϯ"0P\)SJD]:(lE[Y?9ÌZZy6J' t0~z|cyFW$DޅQw,u{Q_JMN^z̡)ǣd_UUqz5Y8[E6-LE{Ck*[JFHqYH1\:H-&x#32x5s~]cVObC6خINW7w<7n~ ӟt6*0} J]B:`^el|""* X ߋ(^iX"O1 3 r_ J#&Ҙ"!Vt;10BB23uS/IN Y$53Gc9tlChZѴiԴְ#(Ŏ5Q;HP_6ax^s/g*JRx KE5,װ^V a5,zSv^kXv^kX(f5,E) a5,װXkXx a5,װ^ a5,װ^kXx a)k4*5,\˰p.eX a2,\Rlxʌ QIjNHGXpi땱{N)20y+MYla6x-Ĩia[W9־ y4$qL`3\&d3%ftnWko)׿azt=3xTDe[- A=G|~u3`886Jӊ*ƂQB8G$R|)⯙cf,#$N\kdX{d`T^dJKB Ў93+Pn4=+x.yȶϔ]O]KuhT]hڽ^5`ńLi>Rke?ޑ^d.vmz_/ ٹ& 2ϽNڇ>ÿB/!6a[÷5~Odyi.M Z0DQwz?).? G~Ͻ~<ѧGP9mI3N)])77ئ 8H4[nR^>2 W0DUdi?W!9&p,#Dӳ0eθФ KH 0&?&|^z5zn0a>V2]3|"1pCFۃyřTr zMw8d&& LSZqT̾;-1KQ'wVM<]#x1b܃o͓ MŠ0짺3^ m˹ÖVȍ)E\S௝!:B@V ᤊY?{WGJOE{z`1=y1:Z.uF{G[sdew$ }8)oٟf=a2K^F{BU+VF#W{\de@dWT,E7M ?pj1o=W,_s*?z2J:d%߾ïmydA}˻WWg|WGIwwz kd g9b1IeX Be*s IO SJ94@0BDkE_޽}IL$q7V˯zKF+|Q0iT2:+ j)Q*eH 3V=>ڊnM?b ݀a ';DKO&yISin!YJo75uv:^_^zT0JbR)8d9*<"iB'#l$l:\BAa@;Szv%Dx,RP%mJYR&hRhcgWj'WC{۶'m.V9uD4 ]_꟯m0}rR5d[4=@wFB!tEalp@] @B Թ`SrI]\iWŕ^j\f>;'yJZ kT4:PByERRi.  a`Qi:%-ja6 2_Ū3CD\ gWsrTZk90ޛ_٤tʼŰO_'isAd>*OIqIfD/.#.MtiN :Թ<:Dž7DP>*L8a3gwhCdF8U?nps3,v'![נ! ݾnϱ'pÙȌ3GQH8Qcig> a8f-vmxLmVe湐Y>P8e%80[ƙ J|Aǡ -g5fens,uV|)6:T*Yya-m ZYW$86@k9#<% MKO|beI!a=W^2{czl׳@Y[[ԺYu2*fYXck3설A6-29+lhh*4\j\Jsh$ol ^!$jd-@FK#[΁s>=_e]J6PFEke+mIWۇn~ŎrkU͖ϲZCRu%Ӂ.JUNw h4dPn$%^ $ &%چ:˂P0\ĵ"Іk:kPF3qE <&)+phP*1]C_rUyZ`pY{E MI#Aڤ\"hu2PCxTEt(:;?OC'N%J/J328 PϕcTT22ja(2N"h;[H.ma;1 !D*E.21yϒu8XVbMbM7vPA5/~>]՛ZJ@h 6L^W10fOPuĔ`ڛm"U^]g1ڟLxu4LosmG/<ߤߚ[+{F#+Gc:7),f]:BKIQ?:_'iɤ=ħѭw]B^]K^-Vi\PRj'|%nb='s?t->*8PrC8bD_p=`dʁs*'4 i%? wTm2>^rin65"vFDm#=MLDY~;킏u\S$)qr)XN@;OS;I>}^Gk,>p!ViG:#SMO5wsoZtڀUq_toWr"B;2@%8H@рXNvOYڒ™ P•F)[!D7LQM9 uJ9 }N8f/CZxm&5[r֥&K<]|^+OW8vLΐ 74%_mğ @Q"(9B3ZcCF/*#͟zeZWuvhu-lץ>[Q`0~fϖm΄STtϴtXqc/x9zA֙6f`u00h8mII'v&hNlke&f3= |7Oȗ :)Y$'([2v7uZwF 6dQL~,brRuоhm,Tt4X,`e6j- Wqg3 )By][~# I:R~ixOǸ+0ߜ U{@f?~m*9FګlE6e0Z;u!i9,CdJAEZNx[Y z{ 5(M3f3/7PmjVP{374)wʈfE,kכjcIߴ7 $1 "P.ebPmwE=Zx'J]mo#7+r ^ 6\W2HLWYՒ,S졁tlYP bAD0o%zǝ'A91Pm%ƲGZRtY+ >g^jʈhh)#)26`pHp:3~֜=fָo^I)Wѧ?}ZuImˬç}ڍ|:P &p@ q˸V(88I/ 2Ӂ H,L]qhh4BkM̨3hQ9b0XJBԀRELf ֜g0e^~3y٧h+g< Oz_Vijj1Gh Jk)5_yr<**-IOkܸO/.As5w1=ԥB->jb^Ao_iK'Jdќf [ZCz͋6 /E8j\>Ir82v3o,Rb~!ZfŖ)zszs聇$^䮂3a( 90aC5͵5%W  &&Sefn̢E1O1xEC(+`hdy}b}gNJw5ASjn<S?f6.tW~psʝ^,&Möݻ%LR{fܤL} =:?rN >VNuC|Gel K In2hҳYtPWjxCg~GeP識])*eފ7Uqln6UwcO7W[}|2(:o;cƌLtn,hFs+%"[`ыa\9vUZ*a 8)FPWAc0bXk(?q0DGt2?VX,@)vGEwo ̂@8UaaR"r\9U4\F4Ҽjf Tlrr3$.f7hu車L.nlc;7Қy/ ᅯVy7GeƆJXE5{#BV Xpi땱*uJѬQ(" LG S\-&x#32x5sH嶌ٚ2nG)b IƁb _-\(}: ׶7]ioamڌ| ,vsC Z\3 JEB:`^eO ) Et!"'UbˋvxM(E.8f \&Ge9AQNJ% S*HcrhؽˢcE^iQzHLdu4(&9bˡcEӊM;h ;R"oӫz MAl5"J83¨.̓% 0OtsBn'—7[͟7.}s~qK>='X{=:3 Nk0Taf[XU2x<=hۮح^ͦuS+9ZXP#z_wKxc 5Ԁ}^Gb4r7>jO4Js?%%:@jFvJÃ&zGi>Ps.`9V/fW2`xOwǎ t{*β1S0fJNmn Μm5jǾ9s7-8`ۻݼ!sñf?5,&N}^7r lLG/Dɔ2 #y%k AJ:^'*V}4 O7vy!4^MTQ%)7?B'-f0RL" x!Z!Pȫ? ~:#h*$_3@TlY+o@vCXtջ+l3(Ɏ:D ^p j~15pGh!TKCD:`hh84߻ޣ5db"¬,X]eNMǡ̓t EǸ/Gq!8RX05sHז[ v+be9&tcp@- ҃yOBe^ `VQR5ƁZ1b"iZ+IpN'Y>tnҐ=, +]~7pcގc'8/mM]5f!K^mjEe`wl?×MmHxH L^DMgw/\Ydځsz ూTTWf ej_V |GY Ǩ5(C(᫄:R$09 &FRStʉK(Cz Y@ c.*5X@U67Qq,tJp]:MPĹ_rE?x0|LDIK[IÓcAGτ {%SkZZS/~,we e{kayaTyjkR+ZZֳ -ɩπ$ t-UV7󗋲2|HUVUw਍Y/ G &H,J1IC*?՜"1tRxy3_ n*5N1XD1`VSo)-'|ԁ8b$FLUQy= lui0ert?/6Pʍ"6mȰ,A[R.k%!RHDcR5g&XzBʹwe[s^G{zp|r^~ĺG1@BeŭLM6h ZDkƅ`|+ sG™8ԧ3#̃ڒg2BL+pPGJD;$^DVA19tQl=N;x[B7zf!% 3Bsa4fƙٚG]kO]z8+߳Nz{{Otum OXBoO Uv\jCusZ{SU(;Ș0ɘ~\cj%l)5ֆ:xgPPJv14QOYѳ?s?WZfz&PBik45AR鄡vi_W+pEge>#? {yT3nykzc0o'YOߧǗ){N)-Yb%>o~r9d[EWv'6XMJ~Sy[pۆ G3zи!>ꈩ㶶C5|-?sW[N]=}GtW[蓻(Gq_p !g[KVXY""=-qZishF99\(z3u% (WKM)qZ],Zbz<˼9{8YPXG❧\ Skqd6o~tkӭӋaY\们ryfAbxd,13ҭHFLDo% #<@u@xÅڍԺf wgTxGDJNcR٤) J3K'V4@2{$P}l*fI5 x|Zʊ[YS A$F9)å 18y&A0tqe]pvI?o&LMO55HJ6":8K /XHcT!%RHyi(}+h0@0#J<3F+"u<3mm+mY'o߼/њЪӷ&-!yBV-hRr?&6X6!0Y}?n,TiƇhNj7"M*Ea2/iV¤y;oc4:B2o9jևL.[̇eeY*D→:҉MCŴl֥x;Y: />>[ /AoyNWw}vpɟW=jk]]ůo+2K_P ׭`=+;ucӖ:?5^BsF>R ?rٟ_=֫AusY!rpvH!y^Ƶ&a@{^<̳/2a-Xz.xNP\-S܌Z~~'H87MYDP^YWҘ/eh))r!8'Yy4i@lӏb4x!+I(%ڙop_5F_'ξkpUu=(Ǿ Y0]2ÑT"^ ˽j!gE8O'=vX5'v8gV>P\H/4I<kttM"lNX5pHS)tI={o։P'@g³! Ӊ3lH'w4`>g>ƒAGaoIpC9hrZ$ ~F ja[gY<R'թc"3C )yR^$"'&}\ń 9U<bT(.R#c&ˉAP I3^8QcQ¹[w23z]2W')IdXt*JY0a֒D[y;vIfܽ}XOMz08U>JVLMvL+oY tD'O>6ލGhSw% 0AX 16&ALT+SSԩ:U;ˤ0y$T(#*ypJ>Ш=X}Zcq1I(E#`j0-:Kh6gbAƩL)e$:`鱶LaXl8;L9oo$q*hnϧL;l =CضYA}:hbT1#u@~T9M UPGrz)'&Q'$7eo!"9*FFC4:᭶&H\Ҟ-QR%Tc[ 7`Ny߹|jf>"h,-vE,cYR l2G^eo}9F7nf[DO|)r-,A9 JTPʩjNjq4QB`;Mil evViUw]Ig"4g@SriӶ-%w(ךZ VO(NDT_a~0-{J8!Y*d![RFr ZrZC Vy"2y/A` v5UFdM62xg9{x#>d VEl>9׻:O?cݍ8CCɪMgwmO?߻n û|Dht6w?ٹ^￐c-AKUeᇅ<4$dcU~kK A5֠p]PR(R~+A5֠[~ko A5-7ko A5֠[~ko A5Vr_RÌ5.nY?Њ?X٣鯓/%c"`QW;wjG#悽K6|܏q3;=?: '8ag/6t) F.,X0K~\bg1؅sWr]-67zY9 o)]v/F ޸Cآvyr{Ә.fM_Ϙ5>5*ìȲ7o.o= 2SlpX`y8\c[[Zb{7#SԽ4> /O 3+cV<c>+ngf'fhb@һjj'ҾAH70 YHۖqFϻh6q^%I'GUҭ. Ɩݧ-sElB!um=cXN}^qoQNVZF>9L-/kx!8ftu{3hvcޮ2kDϚ.zo[^P .^L;eҊ΃(N*}4^T+]Y+CQzp<-tpRlww?" c[`Z&|d&$]d '048LxN(|-jt@L>57yN>vE.ʕ/dyna >cnj..^V]K!?wgrg/J{.f}sNI=_AB;O |2xF4|+Ͻ&W1HTc411 s[ H(`T",bY#Tbn[WUUa@S;k `=(D,ڵsD6$Zj +f@G+U:")Q$wAo뉏ӚDpJa ņϢzYiNt0v)t1:ט6齓 Χ&ME۷hrGuWak&kzj ق7>ZJ [V&43II;%Ay{ˆ@ .ht`ͼ֬dZs8;ZZƠSV ds_Z2Nd܎R" Qƞ* _,\(m~Sl8m6w ξ|ei _7ؒ``0.bJ1 $AʓL,/MJ9$,[Cag9hL*/<(jJ`a čL>&.]p%Šqy9ABv_n)+*|ԂpVgQ {I#^pX*FuTtb闇Qpɂbpl)W"TX%b ae4X*fk",(NS.8yQŢ\p+8p+UNxN{v hCJ{b*.;yksj'dgr$c;ݙ(^MS#c@jĜ{=9ә"7P1r3Ntiݻ.uhmn<:4{)c틨 ƽq5EkjۖYKXr;=[<폱~z{vt՛xRx3Vf >\LѴCBW& (o8sݘZo,&{cUL9c*M5"H͓?f0:&HgO< O{lΠDE>$NB+kMv𖰓]Vű4 k^+$AEl\ #>XX8$2g1~D47H zlrxtsOZww<y˕.B  .+/Bl >Si۽v42'u7Q.X˟>ƒ!P lI'x!> ѣ/v$ [R-j4|zgzp-؟~vܫGB\SʊT7,"'#BC a^-'8xf"1Zfg\IKiLbXEi@LJ:aaBȴQtPe]N#cњ `1$1w5 /}2 Z]pR١F*5uc{h5gp lD_RM×r`%=*)ܫ+:::2X ύ/m@H\9 %@ eF1BB9ݝMSUEFo1!-O 5 cm 7pVF߆ŝI(G빲HLH 4?=Қ>̖\r=GdX8%@:;h3Q#5"Œm(4׋ Lª za:8 epUZ!ՙ ZC0a }MLA$H OE6yS$eU.4*d">)- #XyK&&]*9QG~bBʘNB|b'2*Jx=qh})GTҥf j$4z$fr9C tϩ wf5HA@-|>*<X`Vb6f6~kѦ([kh~it<"TB\HJ9ODtKܑpfrq4׊ g58`5Fȥ$eL(ijȥD 2Q "陒Yv$یLQ&5.pEpkiFP c0,6\Z|Ar߄q.!AmO UvnXj jC.@Lūg«Rb-WWZRt%N8"HD A:q%%C=wDH'VGx=6YGXnS]'R& iHbF)v?{WǑdJl[#"@JXas<%twS2g}_VEMRdI4IUgdd/2#6z[T,DӲpVRtt"6(f:DPp3%I#R!B9ɱDȹe?нܮRrp]r~Jޓ]]}_qM-zvuZ'L˅'38#o՗ &)Q41KJ$IPVSψ~K gR;89[z zS6Ydd0vT5hIYV@ 雿<W;?O+E}|6<,eZYo#ky§D땱e7&[L'M,Qـ\&t&F\rMQ$'BQK-} B{NQ'Ap3 MEfkH*]zdm1"TsJ*Ά\dTYfN 4=hR,kgc&BV&F3_m5Z?둲I m WMp#w~eؤ\d9ZQ/9Z{MpX>?KuUw%E}Ml~? [waPf6U^dJ%D0D/?>6fo:lH\I/_Nqd6; >U~!˓9jnheeqDQZ υ{F6?sn6?u󞏎r6hںK.^-tKFjym:x^TKv PEV8<~YsDUTţx^X~ױ:GaYU^TU$9v}m6 #m &vA >Y'[{y!=vO_D_eΠ`w~vZ}xsqlFb'{ ĮaiFPY=U'c27Q>'3O ~a\-Aٴ7σw9Ŀ6h&#7VѪ\2Z+*Kb%LGc`!EfVA "98kܩ?b#ڨ84qJD2X&W\'*{rɤHX4̗E^>ĝ:CrCvag/>zr>d/됦X]6_߾~%UBZL~K4c-nke[Z&<ⲈD,\`M$C׆WxHRRnܠ~HT*ŅG|N(WBTE"Ykc!/6L1ȹ ,,B>O|}]Wϛ7'V>#TXok] *$v7!a#j#˟sS{*pyt2#/g{OOf{lć'O<ߏ\?"z^Ժ"A|rEFP$`~λ"d&R8PeB97dz|CQ: 9,vg_)d B9,6,u -䇥vȾ02ZCjopr6ѿ(ఏYL8%q"3JȾ,8wW]]z1Hm ?LްuvS;B(q~uw5FWg~_Q,EѓZ:{;^!^߷mhiRO7R* {4v W`v9v!s9bT\,NJ,elM8DƓQ^p CpVN^Xuf6DQ)UL"ۨf;/]0Rj.p AFFTr܆}8IB.(%Q7DdnӑY sg#{1\_H~k9Y<>Y޼]1ukvb}eJݢ[s.jYˉ@*`]Ѥ7$(P۔,"%W%89}mv*z] b%{n ɨR<97k8 hjY{}46>[j ڰ]S}-y^,IݔP]>ܬf\o!uj_={z%3d.VaVVŸed[a7ʰ3N-vU֡mfAf:ކџAvq7[)V+y ɢk0!o_L_wNȿ(g%T;:+G(A/jC[IO=|5WڟO̚/3{d6ٵ}$1F'W_?8GVwuG@0|UVm7wעxs(LjXp^I{$q#j kw|M8| ndB'&PƇDLRTMNaIxbl gW3n cJdra }d ^7+)b̞Zr3aII `9adH8QH,\z6AԂV:'}9NNΦTdb BjSJ HJeCNeN>3VNB$ݧ| YX YT!RR#N"֕g${pDs39>Jiϐy2h`+5PHS1IBBDt0``a46Cm V)S)֔JT1f" fsa:NC a=H fD9F_|q#QID +Wچ@.,w b"v{IU"7,BZ$}.F);2\IQ4Ddbt!`)~b `I5 p5X #d#T3(ƀ(`.pq{ETQ04(go){@_[0PS Ht+6ٽKI ƸE]%FsLם%Ih^68VjD&@Hym$kq/DTPR?{WȑdO.+0w3cgKjJ&%IICR[aR,x񢧰QV  ѩ[0,pZx"?fV`L҄@y3V-jp9&-ZdcVGp|-c pQIbYj9!:nL ݹ`/5`*,dnIKꂄCKe005n8&#q &jKu)t^4DLH2ӲB VJP@aZUh=ɋ ":t@2GM8ڂ6 vESQQ{x*Nd9%$%aН M_/`vsw!Nڐ٘!5"+pmpa,@{%!z]$ ZAޅ\*n FYuE^H.&#) ,C5Jp Ў2vP <@tfC3K:v hs^wŌ: 4XK4BiQi&-J X{0sɺ6녵Ҙ{<_ɭ t mZfirsMLnYƒk%nj!f=0PSm}(8w`)jYu\.X֐D{碦  -ls$ʮF(B0wzB #RDQZ RЃmV#Z276I+dOm7@1\ƒr<@b&r 4B~M^g<2aN؞PҢCHbQyTKF|R^T),A27hâHY>zV19EKkLz֞@jX]P/$byP`ҧ҆t0)I1ȟe _Y#9|=(Ws΋?o\ ZAKrGUqP8kL.+^X@a}W˙-+Z0 5 у =b0#9>d(]sC9ml!& _m:DA] 12 7+&) d$>D )p9X DB .-P] Az3!z cr(jm8u-n<;.5; h[]rpAAE82kڧd_]<"M~mBD⸨ʟf~)矴{Qa}2SKt{PZ|A8'b`-dH$QӜ?ij/zgk7Ɗhl%2!2jbۏPLX"B!@|{]n{Wnmk/ihtsck 0Bm>~io5АA _ %=RN=ăV VBxZI^-kn6qMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMMvM dvtro:C;  :nze힌F9Ze؁Tu䫃c 9pmx ؟G_YdX{#qBƏqUEY4 ymR'5VC7M 𢯊a!/ Юq W{g_FcWz:=s30LħoțTS^Ԝ}Hq^VS6aNl8:sckmPUMxw08 -=Zˇl61rFr8Y6["fyq^ !逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘逘3ua#c_Ak^aZLּäu10(b: %tpe4,]e,y>w`6H!?/5j0KU̴͓p,[Hw< s eL2L2̗3[5Y:za"xf"7ovZ(&ԃBjo| ?@NѶk68)09KʔkrӭEM#f#yiR^^-☡LXvf-<~bcdgaɖs-lL/9bۛQR0J(:RbIppkk6VSw=ZC;znl6zM{ݴdW(;EA~vY;~tFeļ&mu/ŃWM;>#ʽG1TyncεBvp_%s˜tV sUy8kpxe%B?7B'6ϸ㤵i,ezr]ZdIH־ % q ?Xtݢ /JSBެ)[<[xkx/=݉47}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}7}ﻥГVvp먕.}// チNɼ`:&jqAq *7K֪W߸)ƥиts{ 0νqWM\]5i RJ]Aw%Y+ #w{㮚^FKrW[{IURn,Ty>?q)W+D}u|_;?;Mf)n3z|7AKB 靯[!qў=CVM7qU6iݫgPtӪu\/m?Qg^I0"ϋQ 5gq grg6p> GgI]K4^GRgiO~{iA58Ii?|̟7qa\_$ #+rQj>Ǽ;|:9?_ KZlq/;G*a?X_ >:PsJGoLeЭ a:>?Xù#;x+յT5CUGJD[ceJ]~Z8nV#j,~HΉo]pQ^Kuq,j[ggr{9eiK;kh]}uÜ;_f 4BN&W/{=qzן~3mpOfrO>㑒yc(Y֜?{M\c%Ynzړdn,dYsZt\UҢfQlker2gNG^4;u%I*N*!I2,1D1$&j@yqV7|q߽|x0c!v[s_<`laXceȴv<[*sr)HtM u(]{g`n8}ۀs nلAP-Tɠύr*U+Yx3nP%!JÙW:zZ p~zE^Y {o| N&8vU%g"JxO\Ѣ}e`'۰W~veu@ ٗ +\9yfٝ??7O3N5EE٧諍zY]DLڀ蔐08ZY'b lhtƣj|f#glY ;Km6[o^(y=1b"wxoC|6wC#29hؠBCt_}iR"=JF%'obC>|oKت,dA8SLa;N9Ŝtʢ]koG+ Již If)1H"+{ȢH-.VWW9Uևцj+iDXV#DFm8cFָIG~s; cW[B-م]7CK8_Uf>'Ӯ1@rZjBRGf[ƭJGH$D,XTL[ H,L]Q"gv ^>>8'q^^~3mm/]zj5[" hb^@.+:fy"f) l̲*f:`%GT)U'[J(`#l/:.{7pÛ~E*`Iv?؜?~b'ݠ ɨFm~Aa>nQr-n*8kt+7^7Y+oLvOC.=s ܃|oTJJ:)FлWAcF1,5i\p0DGtPYInJ<InlWp9Qi{3J,.X*&2m|r>%a+6 㫝V{,E`LF^)IFum8T[]6׶Vh>R"YZ)jNHGXpi땱{N)2(2fv˸f=m!F-|Qp6lNS_Jmv0vɯ{gnF,Ub)="X0Sl`0BELUāyicD' ECGFӐQk %lR!`$hR>P $ELTvm@.jg3MlNIVS1Q׊E J8 sHʻR+:Sy7#>*0d@zn4H5> 'H `c}йRAنV~ \E$-",bg7B!F"ZIT9/i S)2 KP*Kխ"Q+RwƩQHLd$&;cu9`҄Ij$f.%l gCIi3kifH:ų6YuyeRZW]X[ח^yUuX(.4}8ʅgS;z;eKs_zpe*/Tj`/EE}Q}쎉CiH5_ cчKPKg4@K1NMt~//PﳛEߪT|=97Amn ^uYꅵ{mӗB.~pu~fuzYђ=AodUA;-q>p5>殶K2 *Ղi 'Pϡ4sfv)sYwڢWoNXRߙ܍^ w?|WTnd^mgu*ur՛Hc0͆=]ӫ)d. u==f(L%uw:_j 7M+oMB5 0xiVj]֭3:¿bQ9Q=/,g{ӫۃAw(M[:nScL :m9}irwB̿ZwYObn7?7W3~m3-\w]$];5qh;IQIwע2 aLvzYٷ^Kz 0[cR;ĬW逰(d~sVk/;(>c6al´g*,8]5ζCZAZ!A xx7r\0sn1^rζ$QK-9+R 3]e) | :~(+ÃtU[aՊu8jG0J;Hu"H%V Tu%}70y̗۴JEMit!Q u .K4$q.J%4H:m#b:z*ؤepO)cy s 8M(rneZGEd\w9lVJHH!f咅>I״au\c;Ͻ7ڿf>(u :b b* :U fRS #RNզ2yj(%N~j%J쌸.{&ΔQ^_sB)%BFF_Zy ź}#⦲{Wj3v$ZXbQiţeB=h;*u:h`B3}) 35) '}?&擻T}J; 8nͫk>8{y}bdNGH{3qN˚L 1XTFLe, X<q6  Ι9#~ NZU: Eyُ03Oѭ%64@ `đj֛rQ匊r$ {tJEܐrE`&y FҎYoPʁY5aLF#23lLφ^ i$=WkRU>n'jbti{ -eGsw&< MH1N<2(7"[ɸA!ȃ Tm*d>(]n }ͤ#3TxGDJNcRɜRafXTqXF<*!g1c%P=&6Ro0u2eŧV]_+A!WR"o~֌BqVl|?)FjQĺL0Ɠw^nղy 6"0q8 5-{m}ld|_Vz h8!H/~J unNޢǡqrָٷRdv.k۠iJS|6O};Ǣ7߅۝GP:i_ͤV(aZ 6iR A-Pja+nf(GRe[uf}6_U?\ _NǢ9hT-&~x?hS譽г۞>/:۩VK~KP~We:\|diLnVa+BrLXFF0eθ0NTowa?wwKOT &B |_Oy#1ݣm!#s{XL xZ|gyqQe]\Oo 4M6WI荏Aɡ,X׍Tj鼪ĵ|syʉ6FyRďgx3h>ôEw7MbBVuV)FvS|G$> Ɣ"s)@{N!Q '(PSX$T Rp s/5*+μ%8+7%G'^$D 38<B sư;v4Ul2WK~t*ʮ~әͶ[cV:U]icsUSf+[uJ/'9UQix`,(U2xi23@`b39$]Ղ9+ 3dRDS&Jh PUDJ AeIG"upb }&aOs9ꁊؖa2q NjMBh[Pn1 UsiБ|f[BEEwIsn9&ߨU/026~UZfi6(_P033M>GfY0R VDqSwbXk(h9`|>"ȐiaT~8, KcXXHD0phUPl;Elr"29|zFcO̬h!*逡TptDJaVx^R)ΖfB})n ()&`zK)?oW12#X?{WHO3XG2P.j0=O[mYrKo0%Yeʒ]iL&3Ȍ 2!c*TR5c5rvk($g t҅KsZ$MoYsmE?m@`02~W؊Q01H-)g*RsFMhT@=/kl{NZ͛lDȩ ^ԦP:( 6ER8$*iЦvZ~4 hjq&ZЬWaP+:h dB$"(UVgy9Y>=dD Ћ"i51Ѡx&(DPScFTЩNڹ9aԏx/P+kDiN#n@S\P^x`<JC"'x[DS\jlHۺ)D Ti@8Ifx.zG $"Г栅 vxp5rvk/W ?G8!pZg5.9T/zwzӋ[9 ;x x6w>TؼfÆFNϺ24D #* rN2Hf'`N0Wղ깣)_`B͟Sٖjqh9S9˒9z#c\*KB-f@KkR2D'R(", o#Ą23,KZE^7A%Ä^{GV߮R$AlOXr;lWL 15C4Nrl+xVp*ksI<^7^|s f ]_?% dd~e8g8epi[,މk0Vp.H K' Թ`͏zҜ9ZeOEKj䇒'f:PByERRi.  a`eʣU919ϪcsN͹dW{~NPY[3 uLDn%mV#_f!=ePZiQ-Zl';vm5"f=/hK`Z߃$m;:.݌ A Gv F]]9wU'wk6Ukxnemk<7?onbab]qaqxs5ANDs5Z*+6S"ʣn.fR]mW[/ˌ4T[؄I;'a,PѠFJ9%6*jB6*?tLyM:0ԅ`|v;" vqWd։aT`vRr }+p ֜{*\0ޥSZf'p '8  wA#ڨ%IO2OX-crp߇84ͦ:y \[q&<~1o)<y _] B5Yd4I"_S{(ˢp堉2͉AX.LcR66qYd&7DP>*JIL|c@@2_5NHɸa]4Kuf`*CF%yf1!s5Zi*r# ,ist- [D&bv(f҅rB(t><TtO|ٟ܍Mfp$k[3!u& &NH|`2yCnaH8v+ؑpAǙ3A,GT Bۤ,J!2E9$%6Hg̰x{{q,m _j5U#jA9Tǫ3A+B'I:sd&4/="%xv !݁ x*wQ`~f9;lU+*R Hj5KI:NH4z"3c%_ZmS僐KKqN;W@Җ_>j$7Z2ڮFΞ'|~6{_Z]5+A, \-'˭]n,K={4.g ɜQ'#u(K#tV1: %~s),IB]Yx/0 ZA3%ݍƏ@?kRp Agm:X@h4JoB]AW9 + CğiSɄZXaJ.4̠M ImR.a:(!i<:":M;7;?O Cg^4#HVhǂu_](1*@Sy0'ZM@;A+H$NJdP!D*E.2Dg:d\]pkOpc_Y_TʏNJ?6n"WYaқ=SBfze&wiZɧمe?'wii4l:.6Qd(ao/"_-> sq)9*Z΄LANm.2'qLC|j0zOJ˃O_tn4Je`E)E#?K~nL:ןL97̈́/2ݨ,W I25WUk;G s[o4|q rzS 6hf6_Q,=|9v!i3d_zmuP:/6,M*;mlZ#~F|df鰼o jOpezq^" *taVێODаQ|0e4)VTs/ԜSiwWK#^na&R;=p53锷Q_|~2^QNvn e-~ANs֦ir]\oѐ 4pYaoVu4s;v%ϛxO9(bMlԓA ~2A>CSiDV]NKᰳ}m h)E$5ęQɜJzi繽H8Yw% Qw 5x1PIR&4R v;MuvΪW58ԗ[׼tTmT+.TkpUTu*k46)/ju*O@"0-$gǂ:=( Տgxd߳%}3AK-nNʉBMRh&a:TIGKqbr_QHMFY# _ _\ݐ<ݍaBȾv|8~Wjg29-fh;tT=ё@bZo L "P^ʡ`4e[Dbd6 bkde|nAQ筵zgDoL/*p/FqK19R):h_J٘,Tt4XJy٨R\I]i$ }jP'}Y=HU£a̕JaҹO߼5e-kw| 9xiyo*[G*NqdEZ 9GP,;"j^#kGl|\MvivaP~|-ǜh9^ݿq mD+R+c󻾩#$1 "P.ebк/m̝rşixP*VqJ赌 Cs q]KapDg 2"j+̸EJL*P9{a{:_&on;FhnBlNM,}6 ~=Ц.%ta@76+@GS5=\{] QX,cZ\L28ll7995#IŸM/o MCU&_̐$IQMihLOOMwUS]ULx- Ub8+2u| v{teî,,Uvm&mzvzDKtpş W4V ULv%+=mQ(dw} fycILr*h$a)1!xOYb,$ x  WƁ>1jM8hD,hacOӁd)rUl2_^{Z" NYRC318:FN B &c~AI鏤PbAgGq*"B*V:\HOV>y9+cN3wuK@4ҘMΈTʣjH/OCc I4ũ\8#kBF)$+&㙣N e RIS=E~_bw_uΠ2@98$Xh8$s{OR!0Y)"G]erԪޫLaz%ZBg2ʜBj%}WWJfuՕ߮.Z-+WmʷFQm({᷅Tu4-G>#*J+lf? h[>uFj |4k٨L6}W!kT91g2Br%碮2\]]e*{J F92|U&WsQWH-=5~TgՕBirF |2!fr:u}WWJuՕAⱹtL/E˜)Bn&8o* jfQ- ۊ u4J+"TZ(2l©aͲsyviP4@ujLZ I9h C.$4HFˍi\CI?{Ci*k !4 T%I)WTZ㥶p0\ӆWi EZi(I5]^fx/qI-f.ڒIF˫0FJ@o_=oƣaʩYwT?\FDշ.mG7g2vQ6IVvG3\9VԺ˭e֮Mϥfg^͜ ̆.f^A?8(타|#s.(g0s1tn{ohaݴ;(»UI2=se"`ryϲ$C޾` BB3+.L,}UbK>c@lDZmK,m]j1rr}r\&[ƑUb?$mʼniA-vAQl;6.yt;,dچ'&)%r*R՞ΰM8LxaE&zQwljE}aţy,TX3%ɀE_ꚣow/<'(gÎ鄷㤀o!W^RΕ(!pZ&|&$]$ &8"5j|[m@К˛'&*T{84UGflRKpB]\TliOhưyf| ¦^WSg)䬩^R_&Y2R=Lyd@|*p;"5;xJN QA.ыBOM,jNmR$v$&x뜡2Dy &gsrL6's9*G5U[}4,/Ja[YU :Jk-Eq Ȩ$"Ce&Q0F ʠqA$HS mk T:B j1r j"9q7sl]K_|X)8t9r=wv/nsvo"O.mvok{wjb%.V_u1iNs#DSW\S d x!{yQ(J!$O2T=kBX1$@C TiX x*ta18RR2³҅k-o22p"77%_~9Yt~ӠdX~[R: aTFTSuQ(2!dQp2ڤD+1dQ5OÐ=1βRy NBPж#H CF&Jٯl6Lb⎞hmJ= ؝8ErSQL0EJH+Lr\/|ԂZȈ )yTkB$^ȴQGBBTѨ9amԯΰd[?ՈFF4.$$pʒ&StuNNB$W$V }1ST#M 685I*ʃqy QFI{(nϜ95竈SP/.V,%E^X/A/zq'#D*)p\DN) G$@$j*eH 3xQ zz7w> T؞&ʬ^#zbS?ppn8y2'O5nDӏIA5M*KWZ8A2|NC!UB$7T0[OCK𜳇ң3v21s8|dE?R(A > PRj"#ܙ!LD$4Bh2ã&@Eچ֧LWqt[Da$ZdW0B:-]^+?b ƾ>\>|5mq:<@wۖIE?=XK!^w5Ұf0;w] 3߯/f ذJ~z/I-NL^P _Pi"zu@ {hyHDI'"2L2AMi\ІH  Θ -Q@Sd{Xn#E8BGF3N^Nqq롒`X!`✻ Brtgj@F_Ur~3sHSv26){T rQd޻a p -CH8Br~=)2A ήJ, 8 Q*5UX1t\I/;tR$ a7`0f2/lk-9Ig}GnZlfvt:<ȪNe=y^UV=({ਉ]>36HDJ2%!ABJe9_UUCeJ`:1 05Fz)fD٠32s=f6G}ܕNlK  9fJq!rAZ"Jrb3p!RA=hd6h,twBrvXsH~s˫Ic|@Z3)h &(kC>tq6K5yƥHR(m:gr|*ydm g?.δ3#̃d3!D ,Rϑ49&46,2JQg !=U鄚44T=OG mVgTY>0J0LڝPkG .R,CZy{u+ח6?k滠džlPTۡ]sZ '«#5 >".x]xEJԍ2^Ң x}K$|JVjg0џz'T L:87&IRLeˠfK9pɕgeq!t53/JcT)5KFar`Ff^"YKC,)!2Mg%Ʀx_&//Gi2kgl-W*,8 s˥_Hr!dZV^pW<5̜A!FF[t1%Q| s(¡$!e]b| RɁ}6>&2wd,I60>Mh'S#'dZ }GaY{-Keykޡ 8?j%gՅFYԂSL  R" PXn1R(J pyn^|#Iʅssytj46+1N\wl80*x4@VVNN:r.4DoT >H':gHP)@:e3h3$Q)$c5qL({: BX&RNjmJt)QPGO <(=ڬL4 pYxݾ ɨX,Z]0Y:}>c*j~gY/ FӼ)gyWْ4!#-qE,(Tꖢ$#=037\$aԿNBMo [E9c$fbl~HߑƾvIzCv;/=)HIo)MWȻL?һ1i)m. ջvo/%--;HoB<逾l^b7w=. ڒ/{w/?۳y6/.d ԃz{u?7PZ;}ݸ876s[zV#~nZCqaiz* {Q?gMyo[gc9x3nTXi2.ӲL @O~2?33_oh!Oq]_ptq/VWx}p@:+agnl['S'p/X-} :e4BQZ%fƼ)V-#c2QVM1{P9r$dzRr"1gp&E5a*(P"#U2VgsjX-25[lRlo]v|N>N@2!CE ٚX" #( U qKDAƓǡQTYfr XFt)`FV=JaJ aZD1mb@r2}@H:ind(`3vf5qM9O=I7WS묶J].6uӚI\-GMڝsVͼ!ō,GQX4x x28Be{#%&'tȝ9WEi"Ip-rۑre׶#ax!M2Bwdc 7w( ^Ao.卾)Ln.&$U_hVYk.u OwY&#MN:;TUe}F*Z|Ҫ IIWwZ#7߆U;[XUE/}f|jK+O_V/g?+fo01m_c!PA9kJ~٠m6֧ fƿ\OK>*vQ -egWTd*Eg)B-ړ# \eT]˙v1='+Ggsm(ms[e1֢æ[V<.Qs 'KxP2D"3 >0Θ(T^+fV7$ztJ)/o'1[ܷEOxUh,9׎>sko}.{8<=R'bIc9IPVak9{ޛc"2nXGe-ϗq3 0 Z1or<{Z@F͢""G Y{'A+LCh%%SW$;!.ؠ]B}(]ҍe?CJx/ /G21ZGa WH4w+wnNvy=blrWxd8 4kޑ)|Ē<-U~p>na?g`p!jெ/ŗgMnw|!T7"5|ˆa&.#߿av\i%61MI/z4$YR8YwwY}ネݬy}З(5ޕFr#ٿh,3S}@c3=`7 r:}udIu+K)Zy0cXml9y+y,_-BR)7էdUzKۯ E#+ |W ẁ<<'%ٛʕ""h/Ox\@/x SP:?8땒D~C , Kr-c1x_C\*#[ҹVtv KzJ|A[Ľ!OgYknYEl`k.PHx3jg>HJZ+IjoMىk&r߯ jzfIņl0JJ>݄,Z﮲lmhɶ/pyyM?EEmz7ODh{\9 |?o}._%` t?__4q"]1hKQbޛ L/c$Eg!yC_gFi3\X".YK=[D5lݜd'a ✄\3O$s$(@ۋXii\HBA60]ȺE6ɩ\5~@2f|S|>N>] N; `?]pᅖ˧kA<-nkk>~۶bw5Wᇻ(nyF;[8dZks^ afaɰ'/]C0ew>.,6mNۆ}~MwMpr0vj:*6>ɩDc3۟˺ -jKj&L1F: lp\2sVk/;Q,} *mԳ ENLm07nG2VwE6Lu!MJ4h6v!\0s,K{ɁmI 9Zr;Afxt=0*褽ZV ԪNګVݓNGM4he*T TpYU*?g )oh-f>ܦU*jr̥c bS:ι%GsQ*)Fii[Sa&poL,\JPs JqTQC&J#"Sw9lVJH  ~`|i4v)lFo'썂XJCS1/brݝ`1V#Ч=}kƓ#%>s t D^)W_ɣ`##, Z }n*$6ϴ:O]% ?oi$,2BL+*JD;Ƀ xuW Pr‭i_"bK(F[ :aFhT5t/?tۻbh>rۻ[|-+|:nv-ֱtw/ݮ+0U܅ӭ^hV"yNETRSJFS_ੈDrBtB/H:)?6`Q196``A÷|(qd D|g@y=t=Ye31ck*iNYV j0đj7JuJ9\(zaZrE؄ A2 ǨU1kX | )5K5ŁxRFד<*J@Bw9LoN⫵}ة ?tkы2[eD3wY!霨20,HR%sF2 dHV2nP(f)[+r7݋{L^nƫ0*#R"%1dNF03,*͸Jb,aDzbU0$"XTZ&d>`ŗ `sSKbL)n6 aʺ707ުn>~ÂWQ ?'ſ<.۔/`Y?#h_$ś0k!{Vl|f6f )bŅe'7~yda\;3r6Dao9?ؖl=\"Kl& @mj]y>zЇϯQq19,!N !5qo ˃Z0*S%(Y{'؟WnyPh,_ wf?-e ^( 8H#F^M˼JNy+odaSTh\Ɨ,͢Er%~l\A`3w7+xi|.}q&-W{Szp9>{2lqn>C!+fJ^o"kg5 YX7aMmq)75{Ye沆@cYZ=DW_eѴ|Ck|V&Ls)RJVo\oFE@ٴ?(* q&fr>vf6ݛ+r*?310<6mף*+-3ƣ)wJ@KtA1԰,6;Ӝ|M>ǂϜ)=6'ۯ_Q=+'wu==aD.{àB;~fZZt! 1HY|K.֪5%ٽ\BKQ-Β{\ئXUZm:]%vtJ"JY{*-tƯ'Ӯ^"]Iɐ\+w騗9&kNOq=L 5ECN&ˢT0d3֤5 p9fmτgBqǟ/?ƲU`EXk*}QtО<]JMEGW/Ti1 3ZCW NV5R ]^^!@Szic8]C{ŴPj,GЕԮ')J[DW BWD7ʦeI"Y+UB+P*HWaMss$3LzҌM[Cz7È7`(B4542tN(h4TR"GLp/#zZx^ ] ,T`j ]%5t%D7JiW/$MƠ$0Jph ]%4%żHWRVM^p86"('%R OMEtn]%OW]%[ д0T,'ǿMlL͢؟_{* W#6>D Js)5L}/|Q 0zcկvA;"F\?gLҋL+FwCx~8 zLJ˦_VnK]u!R+DsXI%OM6L4 IGAj 3 ~on`[Ea:~ڸzu:og1(Ig̃m s+gcFPn Ft[mfKZ[[?3 G.ȮP#.].z]`T+,DDꥦ0+Cʘ:Jm/ 'cꗲ}]TIJ5.bRJR׽)Ҥ+;h!Ii]vrl8nua_ F Dq4݆oӷbv.|g7rsoY t|o(0$wf1yZZz9o|45?-fIcz;[o@ь'5^_k.֘]0-zn,~O!W?9]` "BEULA*o#1Z k!#@nPga$s1# Aڈ,AQ-uD!c k,Rg2DOL qg(%5Cmpz$ bxzsip7O]q?TͥO.&n>J ?n#gNsI0Bπ<_J61j#>:"& Z7Hdn$iGACp|fG $Eec@^Z ZI:A ĴiDF}Q^XzZl!]B݋nX0=,ėJx}Ou3 qBE@0o4r24 Y*R[J8xCDOFN{@y,68&Իc^٬Y7K%]mR$v&x뜡U<j[#T8{a ǴbqVVyn*W}v48Zk:Jk-Ecq $LJCe&Q0@@AcAHG4ǿ  &@)E/4zp@jPi 1= FC8@' SF3$ya2ZP@ != &D%L{1 /X*FuTb֨_d`<[J?EDV逈"ޥĔ.w@Y҄x"p ''!3EhJTqI9!DPReTGKVQ9raD,&nDpbv?bZr(.¸\pNGTCS8;$Q8^GnLʐfp1phvP0<|Յsb3v(Z=_g4+Q!g<rUo[!ɡ2}Zٷ*!w:>IعKX%(sJJMdz; D)Q P)1̰nH5EqLogq, Тd[7wj\[2vRUןFn9oo1}7}nawY?nMO?XE>5X;[kPy\UTe 1o~zCUGt97i(jk1ԚiD"Zփd' YG(7@N c 0AT.%l/u(T#.F)!tD1eDzr}]l˔4~޿r˳k!KE/l&BS`)ss-SpuW`rmmORljjOj^.pYFɻa>Qm{_};*6Ϧ,lXR/]rg7]ߍ| ݙrlgAJF"ZYwx?RFGj'\"յ-dzcHccX1[ԮqZ|}qZ>ckZցgK"p𷢉m)L>G'JHBI fʞ\J\h&hSH࠰;m<8v(Cg&NB^XsaGʰ׃mΒ}S+|3myy{?3WAr-rF{ʽbd|P8&AcA$C RaIL$#vײR|O.b]+eLȔq*!1jSBPce'/`qnrnPgQjꇩ驦se76%׹;&-/RAqMPqQIDa HG9IJaGJ\휼.&E29G3!rg55U?XtպjY׉Єp\y]dN7mUC8}MH؛j͚yNM^NDUꊼZ^g^Yni*߭}]FrܚL$n5YJB}7!~|Eڋc-+8Iu3>SYnT.MD]эQ|E+Jrsv{z[C׷_7e[ۻJIx J]#hy躗Xm{Q⾕ۓj6~ޠJ bUofQmkktn}Lzfo߰](m.aa;Xn}4׆+T۱_~ی;ba~5ʶShrq~rRMѪB#pr1iN^ū]DpbH\Un4>F.a36XfRoș~8/ŷ5`ݶZoio&5/^Ҵ[^l[ox7{Czp5>Gvs4ތf#hle=zXziY^l]ODbyhI$Y5AV-Pi#oGxĿhV~?7tz1hLp9_m: yU]vūlOWX?P+V^gU?wFGKűG^޼P Z3ZE*ʕ j]Qm$ d8eCCRplW<_V>F1V{H4ԓNeF\{d)pTلfYpAb"&Θ MQ@\Y,Vjb8sdi-({4vtM*"H.i\a%e8R e9k$r|M350/*9uEZ=g(+]CG#hC{)ۢv_@46̏.gD{p}:Ly$J Lsb~1)eήN,8 Qj5˧j3b ]_*g{,ZUЪ9}SH(b-J5jR":MaaxfLsl<{$y,y&`5`MGr2xL{y)j)L`T$dgzɑ_{ vQ5=h1XL< ^"Ҳo*⪽ٜU5z7g#'ۜ7g)q7DorD_9I^Hw22a ekŤRLE"HR @zHY:` UE mPAWiόd*$V8}](f{õhLĶ ɘ˖貌Nv轢Q0$bhQ8CYI_XP)gSlP8/B\ovۏUCBϷmYBO lHtMI/\JԑqȤG76+G35=\8Jl^*o=9P:pI3'nrr52(Ȩr`Ĺ݂G󼩾gWGsCІ,o+"K{6͖".H+oy.R<-F1#wUV7ž"]wW$%gйW讬J=rWP6qWE\d⮊*RZ޹޾eʨ% O\'sIin+{we;wСR%qʃGuAPS i;] ㎊:%}oK#GS'àLu4:SңGb[j5{YI`x"YIZI=keY_gF' Hs?+a_Uۻ"QzJHJ?JDPȩJqE' nɩiuJRj9WTR9!էj}TNkxt{T SW4-O.+\S␉޿3ƣH9~GV+P=C^֖ ŷn 쑛&AȽqE\⦋}dnHiD_=rWE`?侸""՝z]GV\*⊽ rURuK̯]i_2l')xRԒ>~Sk1Ųd mt9m3mvңM0 |lYs!>C}`F9tJH0[ƙs;Μ]:RPR{V|qNʼP|7֧x?jP UW@LW Jl)љ$_j[VcNUQΕO`*g({|'Ҍz°8}.:ՕǫuIlԧyӲ44WPlqcSԋ9O+?6Y3 aò^1LXBLCX-%ZdLp&;!Y(D"1XaSZEL Rkp), q(3{ $.GMF"[\]Yjg}D{MaL5;FO8I\MGC`4:.̿UHGtf߮:ࢶ>Ni>59?~}Odr,AMB4fe?t|MAÌEpXdixKӏ{a'Re6{;Nv6GcRT3oFJ Ktr|qG_[|- }mC#X6~'Y{>HSۗ խS}+PrRUQe, w$}e^&1Y..SK]!갂KO_.R]T?/Ώޯxͻ]8LMnv5=,Qqֵ^|!#8z|i0}4cHpY؍|1kWu2]qǵVxo|5vOo>*;.KܔKY\N/Ŀ1h[͕ j>r}m 4_#S (8aIHk$:wLm2g>^rEnOUc~OPQp4 ivGM1)G^*I29RN(%"d[ijgi/ζT1.њNd?s B-! Ҵ%w6M;^O5_XOuѥj ꊪJ+* l?z G繅*+LΘ '*P"!0-w[ԑ=AaO=s떵_ g.9SXm tRN!Bw!l&40Q{.{"128hWFL1`FE+=hB)2NjM;Կ_qH/d"\^эN9#>IES׏Z}F0tM&=:F(h(ѐT~9t^}Q.Q*!D!簊0RgֺG2m..t굊4CJCY۔߮JoG; [䲒`UW$H >ާ`Za+3Wb3o{oS'4ȦoÓpC7jwUqLt1~ Hnh{`vR¨e0MqܚR*[DOzeJܧ9 =#z"-y"v􌯇тfZW/(DA>*t&'*/0deTWA1G*#F^*p/FqRLT ڗTY1 bD2n-(WqS gBۄϭs{!D L_r=\1;ώљ;&o>} ?!B@0P!/+m#ILG}b4ߦS"5nbFu0I XYɬ/ȌEguF5UZc&,D69Gѳ[evy{Ay_n!Ҧr&zbX:޾F|T-҅ Z#qpdLe "I6 ,kXZ]fLƁQpj?}#cZ-:Ĭ\hި| *Dl 9N7Rj.GdjF2- eHYr2 tIRe>q:pX?Vg(ԩw!#WN+t['nvtz7b`f$0R,@LX/]Jdԑ ˁҪ{Y8h$r Jݾ U*rwI Upal99n*hhR\\M3Y drW[v/ֵ&bwMZk޲F6`a #jV|^TJ팍qM4iUidTBt~GT *]DLsٚyK9pɕg,g2jN2dP%"9I1:,0#:( z@b)OL@%+U.O$kH.R$]MFp{ '1iN1`#* \H09H#K^U,3xf gk40!@AȢH@U2-I'Ձ!^#3 q9>p](yVrbl3}|̒ٹ"EUyTrὤ;ϫ@q12rzIsY ˁLIM,,avaG/-`4A@ke;/',޹dJuߩ>+ }fuPG##-\(rd]rś [R). KJ?$:C*h Yxo00k{Vg3zBT~r}+i73̫ N+h/^8+`}QT>mG}=*SBg۸Y0;zZL;M7lűLht=yxt$mX̟{ ߟ'L%C.BNxpFq>㌉rNK@A=A.Jāw7\̏ЯEWo"Deog 1y$;/vǤmF2u-hBe^n"x#Ar=0 A$O6&}fAd%ff.kFP+7Ĉ>XJG XFy͈ɇSj.}Dtn2' *@3͋V2$I֞MMγv~ ]tkF}G;1R&BA\t5}N|3anqI bCva>cpE-x6t0TL]yA&PIvH靨'tсz_ШPUfD4.ΕÍ̚t]!6eQLUQo.)cP닉;m?lR{uDIQSy92d"Ì֛hcTB%(A>f4Vȶ}nRr)ie[)E$L {\b M2Gw%Sx+2w-dӖgSmlф+Iw.(9ޣM9ǣ1BrN Eu!ON`怩DM f@ƐT~W龁& ɦWlXH:ind!%-Hg8Cbٌ_.jO8[^g5-ee\=.VIgWޕGM;<3y`d7f,GQ }hz\<. Ҏ]*!솇 1vܸx'=wh}p띵6J]L&Go9<~!UNPٚ!7dc& UM_j;f JڥP"{6HoT(>A2Z[@#OchNd- ZIDKVf\I`WQUa*1ic\w^?.޹h((`4)@_]Y+КSKMb}ɗ vWdsWieS/fѲD-wt\]dI ӽ~|_NWY7סwN\U6솭xۯ=FTO.8 ŝŶxyhRl0&#Zj1O{tql]7Ozo\# 祖a:oqusN|LϮ G.'{6 Ts;^47ۦәrmocsΌ'3dr搸SəSs)s9KDqBp%80iOJ \iѧ*RZ+I,">E\+NHZH }F7 WhOd.R!KM2>0d Ɉ{a0`ꃒџ`:74d;a0%B{F}{.)"0iW2Li.R.=Lzɟz8jq}0s'7.|yr3iG~Li#udv= ;>79aExBW3xH_=I%mv Pö`˴PnϸSW:'Twk0Y# 8 R`Keq~7 QJjXXkKB{ #-޿Ҫy1ŃVh@ؠQkEɷ,!J МECs8":Hḵ[4EJE--&  lx!VS+:v°\!iʣ'qm|rgnFmfQ?4MrP܋0*39 ,m5]PI**OWCdp{:4HkqHٻKM\EtMkmph ?rYQK x^|MlU_V\& R65Td*E=[v=Fizs#IŸ7$w  ~FJ\Ӥ¡(zHICE# 4S]SUZ;<]H:~jj_Fu@s7U)i &$Na xxxOI.JT{B7V5e[9<^Ͻn0S͘XA-uҹO>su=ɝ{~LX~7:oڠ9袉y}p6@JPMK1  Aփy1f&e0ᶥ=0"ALjM<(ÁDVHu mm"9b<=wW8yLJ 15bUnWI gUw<7EKp\0rq- Ɂdʙ(u2)r J&Б'7S! BX PoȀGAK5a'<s8FF#&tZ@A,g eCPKH!;%fg'g-gw*4Ь% YPL0uB|Uw`ܥjdG\y#_,?:ID9;>j{U+ !<Yh^Si%kC4Q=^QDxl脞y6hmb G014g>:"& Z7Hdn$`O Ϝmǫ'Fg?d$-l(2MOYR($ H'0ꋍaw mj1f!}B=n,ėJxBw(_z#]*8òPQ8ܙ$ hcbz-%!srip5?ڎ$܌.4LZFѯ#2 KeQtӧJP>-kM0׸HMVx=Z_Zoxks~ֳ+vWU?x?琚/2=^ĉEwfRwpJЏvի_nUtRD߅ucޭ}]zV~mvs_@9o˽9jbKw;)zԒCK/v?,+~E}dhőEXt{Bo._Rro. YeU`IGfueX&KY#KN>uRň8mhN}ꪞ$h*APifPHCo4rS=<R>P4@ujLZ 4O :$9&`hQ-u̅df2Ɨ٧d)s/%ٷN:e֫)ڃZKRSN@E %914ZfDF ""i  ^O}e\k҉"?PcSъkו]^ʢZ1=9tQ^T;q9r6fvA\4r*455 ~J;o.}UŔE!ZUKI8%qS)=ؠ!(C4Z $J!$Hb!5'NBOk99l :%*I%r**PB*ُJ1,,b o[\efܱ+l_Vnb׵_Lo Tv2}Wؒ9%2"LI( 0H2YhfmR"TTejr`3dcFs \*/IvEaHc"^ˆ]LfӰabζp!8>4ŀ ՠbE{TDqP{ %O$v m9dgI$d622C hE{bMK#b^.$dUꨬ-ُQr x2<~0"DܥĔ.w@Y҄x"p1=''!3EhJTqI9YPReTGKVQ9ieaD,&~D|b:rJkŴP\q 8NGTCS8;$Q811F1!(CRjca)80axxs=̑{W63n(Np9k n$~|LG{CI^?xC yHڟw҃u*!:~HE?JX%(sJJMdz; D)Q Pe4Z>4v 4c\k/E8#m]MfI|u:Z`O;]==?lqm;v}*ycw9 $4dve3߯kV تJ57 ?P'zhod>Af: 8Zmx J3 ֛ZÉo5>\tN$K- PL5AKuI"K`rŧ&٥\#N3Qf0oj}5V٦fj=;Y^(ZUhd.^-&K߫ԇx&X8񝿅>Fy/b,3үؤ] q_W7[_|ϻLuk4=-CKyso4m a9Sf2rHYqF=;b!7&s͔7F|ggg,˾}VVdV;/uB=e|CO$nݴlb_M[HZ1o{qim?$=-7К Hs*-bq8! '-%+qprK i˳fu~'6+Ǟ НA+YL"ZFb   ȰTfӉR3(d]AAo:q6w?խTI 1:i5}5Tˊ++"@պ rQI Lqʆ9}g-s\KiY & k0kd \U6{Y)\PK. 1AxG+ͭ8=o)T/b>8¹:pCnv~2cΗE;hvq'ǁ=;bcGalϧĨk.!HEg09G 8k$r*F)xfUEVS>Dezq*FІ8S١v_Aʲ41O.hDENz"m >73R nlrOY%4xE{v/ywIvuygzh"$%Y*EJ@ :#x"aWfb4@89WUٿ?a#i9ەv 6ytǨdp,ݔG{YY#ęVS㫣7p*Xá8 ϭɥ˘l_;y-o^ͽ/z6Lݼ]PuwXL\qN_0٢{6|A>g7V2"uxszp1z>\]Gz 09!Ar<<5zn&]Xd%pJ^kfE4 lKBbGP m*^A7bw`aWF;yH0rW` "9G!RmiWVW -`j\@ D}RZN%(N'L \ ͕rQGSw 2΅H/& (Q9SRXÙs&ZrBTjqGh;YzggFخqZϹG 1w1F:Ǚ'b&l'CU5!R&?c*JA,(Vjg ?~jgFrdie$yF,j/$ˍSK AJS ĩv!ă2LXZT"=&bCg}y>qB'zwDͮ{۪ZNږ]_P9ѭ=B hQim)@)@8I @t.f5, T~oHP*FVG#$K3J1&*ybĹ_WPy.Jv~]VuqEY(ĸRl~NRNn VJ?%z4V ULᖶl1~oQ"h(z,z!&CgQcFZ+܆ 쏦PbAgGu*"B*V:\6Bd҂HIwd}ri".Qdm2,D:##{N)W-G~C\H)u^\S2J&YiML4uR(jultl2p]{ˁP[fdj-nU{>+gc 8i"Y ȒiKĕ:V1'y%l[,ƽU**oܲ3( s%bBӠx@rbt)e(#IMB@C"o , ʻG_] (&Y/ui3: Vg+adsCW.35 ĥTL$АA3CMP3@p&e6Gb% APc&/dRrenMs +̌<1o_|@fNϗA*}NB0ޙL'Чr& 1!hJRbVyQ z{9y%vҚjix=K1xf .Vm|sW|Cp ~;%*|>N\NwKT8i솨qRJ_DD%*|S RB1WY\ bB)5ez7Zb̘<sUPJK(\e)Yk^9QZʛ(=sePUԾ,œm+2WF1 n蹐T.':jnaw4_",ſ)~b8חQRZR]PUwh.L+4H0S Lgq:3f:KZTʹ ĀJv0*K fUR\Bs%>1dp,3,16dfѕ 5W\)IdP``UW QRٚWh4@ 9e` 9/4D"bP2Y:|;_wAcXXƒ Kf/g}W+<o~|Y(~7-WZpT,$ɨ̻=aDD&P^XcuSup2I-cQ[sty\j˻nX=S#tۛ_k9ۻ}[+ :!GOKO=s

gvQ`sѯr2:ǕivקED6n!(\=hiA^Ǔ"qxi(VWG(WսkVz>.` Kq 2d Ic.+4!ɕ{o,zl~W_#]=VK\[Qa펗um{'W>:m>_{aٳ@]=Ⴃ]o;60Q`sք^ &A@c6`BsdVaIL$#6ܣsI sw1ŭoy1fJFɃLU4Š2iojEB@x* UBc52ϔ*Er=RVke}JH2NFJǔ(T8U>ye&(R9~]ʣj?Ob8ē^]~T6YU5>㗡1]ʩ5Id[W q\'IjA6z+_[V 8m~ڞ:,hb>u2-AzSvN?L* ƽ[N=CDԼMX݉~W$ye8TuoEBdrRWDi5!_у.̘;oumg9)kqyө\Yf -vS0Y& 6fe]%׮ֳ󤚮:^]v}%X@,TIn1h3zѯM⺅qwml¡@uhLp诽1yƸ.ra F%,^Ĉ (%s/t5*+s!k#ODyH.()xQQ*e$BP𐀴_4gzȑ_򴻘vx)dv{9OXYHr'?VKd$˴-; MU٬b}bӷ@mĚ`% Ÿ>L{\tTmxNx-vUT8AUQTuU˖¬ V** U&HUE4 @]UPB0GN)BJ4f<.r*$,%4DW{wGhQϓdJHqŜ+PrhS?8h-l +~oƣA_f|NB$A@,.5 = qAZ:FAW2 IH %J80a,q{ ,(4X%JG Zzb@-ꦫJ "}$9y'LZC@4M`TYgD?Cry\uE)cDS|QRgDpMUV5Jkbx樓@ȠT'c$cGD-g.ka[p3s>;P+P̰'SՄ>/[&NKXW;,bТudz;,S6PȚodŴX*gHŸe3Cդ7ZX;|\rX#4(B܆dJw0J2*jIbv!Bȓ]j3s'"IOK),PIrn_;VMvFߛR^nZ}|0b#uUXArJR~F*?nr/ui3:㔠Whɨ4h )!)\.NU >P1@C. jR 3),8,Y֗6B8=ÏW}Q>NPX{qX KRzxp2R[8ޓَLQ#C,A^Y>[e^lRG䄣"T $1T@D JFG\3r 1CQaM R@Uh\uJˇ^J.cac;|^A巅| n<^ln7׈WIO[7[ϗMJH# JX.G:N&EAZɄ:$\N<Tt^4B4CA1HCd%P1p575(4r0y*n:t!EB4.gX~/wgq??Nb7&ң#[5x'x7,2\yc0(TfCe\E$O3Cg|N<7#(ъAZy}t/E"D3AM@&$1gR)yn NO7 ZQ/YsneFv}AlLJ(,M$F5qUQ[P!p4') (렉!s8:5~{5~L BǼh;X>8f)@/y g:~7zZXb6͑k%kln~Ϯ޼%iO)]`!BO-/ƃJիzhs9WklٛͿֹf_-/g_$%g3qFY\h2{Wߖ4aWd7;yf*;s}vMaqMǣF&8ٿ>.sĵk5]W.]"y]7XԷo73`sG-Iwח\Y{ޢwmh\99T.{yɹ"꾶~6lΚ7V>OQ!Jp6B]뵺X`j`͢u:U/שk7~!ϯ 6^-nu;/Jet7/_gTO^Fқ~IoOv`t]V[' ^VOʀ?zj\Blv;D~^N8/HivȱM#Γrz&dr-'';UC{;KMedz<@ƩzbGq~l8Cst1v 5 %?W?\o oC1+|iM";MjǃN5[ϝnfϺ_M3?rsA.0"l sMB@ D'p*c ņgLE=X2[/P@D@) K>򬜤+oQJC@eB!I9'G(ԅIk(, } rEs̖kYrcv덓cľ1%bAuΊr29Cq*׭f\>r۱6̜V/v:_u_ ao>>j:? \GI+BE`F`i _9GR>P4@ujLZ fNH+ AR\HFi,\hc"1M 3[ |8^%S69kz 'c0>/΁+vSrYpnyp*(ɾG tzgWSԃZKRSr0zhIL,+2% E? ?|>C!At"8;5դ1ۿ.P?:ZyrSv]h6mٽ_}}BpNjpo%7T;ݲ}sSɱST+W`]Lyin"qJJ)R('{A[CPh.OjBH-,.h?qPD -'ǂAV D JJkb׌lb8R66:]Ttᒣܷe[ve7EX;7AeѧA:{[R: B ¨$PSQad(6)J * p2س,M.$LP AVI`d1tۥٯhV l 'MiƂՠ)xTDQ@ %O$+Lb9\y2.I -C = uMK GQxЪ˸cg>A"(,+ƓcK͏c5"+i;.9)1 \: AEC:gJz3>)> )HNMY4qy QFI{D gCaXJG|tq|fɱz֋Ӌ^9qD*)p\DN) G$@$r7&Tʐf}œчfDZ C8Nރ ss~.^#z6S\XQ 낵)X=G?^0s^P _Pi"zs؅9vRZ^,:)PQ#B$=h,9xB*!qbd 0aϠ<bӉ.VFn6lN6QDoh4M\?@ѩ[̿H\v+&()PޙnY9Ǧ4moj>/Þ'}/︚So47 u.}2;KRlOGH)6K$TOE9(R gD3:ʆOc>~d3yO{utxubk/X\>mq=n=ǹFJx-ls4khe:%* *7F+BuQӰsm{j|u&θ FyYOq؛a?|E)Zuu8iyԱkel_N#}>UZս_;lo z_l  +6<=*\:m |{i2b} 5d$g{?U!ܥ;FRxUPgJ82x#}v'/kI[0SZVCҐ5({đfzjj~]]]u'inp IZ^=Q8g\krG]˻X ncwi-^;NƳu6ȹ a31#(hQ' K'(`T9F$㑅H>n7b FрG!eLD]rvl/ƣOŤtHge̾ecz9;s|C|Liu*b5Wx+nX+}q zd a/d:`xBՠA@0P"I0Rt%qc#IǽfڨF;yVL}4ZFEjhR 1T:mg!鬑BS{=D9J$>=>&)xhٛ7~{p~Qx_9.UGޕkl&'-:9zCorDs,樾,{=O{%ƾ9S`w3>3y=_|YP(H@^FE1E՛k3.. B)ʓ`ypǺ6_7 T!!3eQKCOLʚgERZ\n=i-ߕwJjp,U9f( U&-2͉AVMJ1EӁ2MMֲnb5nM~5jg.JK6#6"b)S&r<#3JE]dVR X⌉1Rǐ],h;[__,hK-u}sT:Z=@uQ@)#扷,aF <V@/I`:&Nxu Z]k@]g(\%s"t9I('=_Wd9GTp+ fR fh25 >;~r.GlQu2*@JN'7X4#UZ(?<p=)?I f~17]чK ]V3e0;?Կ_x7n鞥.t7.$LS" E!lKI s`y<"I/|  Eq(̐_ DwVMWۖ3GnyV{{r0SPgKW/OFLV _PT}7+qI =xAחYyNs6..d`/\t|>-??@lX1 <*S㿞|}5&DW HLμ3y !uE MȾ* ^YJVR+%~ <\۽СC9~ ]͵J/9hw%.MȥJ=ҪlŃ.OfR0+USY_=(0 tZEǸOqڂc`)"<Ŝ9s˭#i+XۥxNޭFŰKQV?1*&&VQ0/($<3@c 4‚Sڊk~\M*MM{sO~"# `f}< l ҴPleWw?Æ}M %vF,=4#6ي.Zt.Z&OG|n`3W :#"d5`] 3o3KU < .wѰWYO S_̓I?dYYbQKY$a߳*Zf b<+8|UOk^jȷqX^7 b\l7W彎޺ Yl=k- vu]B;0{YvC"q:OӜrƂQ,G8G$tɾdkgh{ӶNLt`ךy"֞#(+- )!k7yC)Ѭ폸oyA3mV,="Lem/LZ)؜eud$u^n:vS{Z]}Lv]^hFR0ЫnfS2U:o Mf)`K_MY~t=;7߶h3ѭn'k;X""FG9GVҨrFEC9ri1T3Ԫu+¥&LI8FҎY쥔7weOo;Tkl%61ͥ8I0BOFkZool~+,Q[ۥTJwe;'Uz^e`Y&'KpEeɈ Fo%"*Pa6mCI/_u0?*yPˈx`RHitL*"Q* NeT$4 Ll.{V JVdu@ఔ?n# HK  *D4RlEǕzs?O.kjF4;m"3Eu*q^8r/B 8^KFwH"H B2 N v$RO8 T73Dx8d-Wܜbp}5f?d"n8xմBR>}WӁ6Xی0\%9'֌CvVl|f6f5-X69/+-üO/2lRga#a H߲_L9K/8]0j7)Zy®|SçcT\Ls'aItO F k܇Sdyp~H$O&0>ޚܙuL|7fZtOGiR67X 8H[hQU4!z=FmwR*lna͂UT@ظgF3ғ:@Z,NfS?8Nn|K۾}BGB(_nt޷bjc5\قP1zR'KS Z}72MW1Rtd:,YRť¿gHbox7 n0 , xqk^GZ;(. A\?{b.Q{ &\ tSqT+m#IE&<"23w``>,Fnäl{}#DixIe*VEeF8h*4o6֧lCwuWcnxd1. t ~r[jr9mTҎח0H%mLW@m՞8s5eD2 [.&M2LNd1HeA|2$Qˁz rD7|8MO8Lfj%vL- [vq.53R` J%hYjDbpIӤ|>#%S5^=^O͟ \58Uǩ+`OZfcIѬ%<k̐V ЄִU|JM)Ը "DHM4UߋWw.űqשD1|2 [Vo-] Wl0)eD*8bCĨf7xQJb$HT58> Uo3aTIU>Lz3gp! kL<>|)τؘ:h]^R-._22g:鶆YtU] MO!xTnt2!*1& d#*Z*Psn@FJm~GKXcН@tHw~I{¨y4B{p=^l'Ю$VIM⮧Ļ 6J[jA gnﻩ^S BjGo,.jBnBK OP%Vz깁&2a nh z9%W 2m9 i+ٍA0$3췽0Do#`[{m{k3V:?h 1X'nam T E덚CXcTL&'\D eH6$n;m%W*:%KʨT1[\L (TJɰTQic0$CdYcsn VU *^2,q}m?57xi_KPiɚ&zz?^A:.vk`e}SL.ҡN;M1{%7yQ#xEӂux, L4A yz˽Tp@`' eb2ӐdIQ&f6*Yaӻ(N~(3NG(d8@*喫%gPCc6}HBFgafぷrsfjՖDF$HHL%\-)AE(Fs= H9dP<oYq7LbsP=!'w'p_nFI$\xm;pٴ*w˴ $\o'We)19wM?c1HS>6*^ȶ݂{i Lvr\٠u4eͥ ~O'kmƊg=q5;zW=?ZcP($1uL+~V^<6ZyV^7ۼeIֶQˢCney n! vm(AСs['w܀ <|[&Ttb]"MOF otwBskO~lړǦ=qQ{rWsB2TYH6̤ GoNcRUdaqFo[iS<1Z.;NO' a6ȸZ;ʿt줣^i_O4m- F[Z3̱5k:tfBގ[RtZ)ejjLWof:&fG3]uұLWZU0]K#1MWl.׾f>Y Чf%0]U__?_]]Y\Yw;kkV봟-f~kW~7 (kJIZ*hEgj}r nTNZx* :o %~_L&e\USa\F,"dY*mX^jtzjDwSW,uOY]s)؜_&UyN6f1ՈQaY{%#$1K,\?lQPu,vz=o쫿΁:.XǛ߷"CP, /m6 *ARB&Liygj2nwi{9^?Pjq; 7m Fٮ{%ݓXox+{E p\|]^0\H|:$J^{2D(C!ԐxPO9kK$Q2^]4ӆJ R>zEĜBx~QʜUňRCSh]ԶhYSr:7˛뫳eA[LDFբ#[mIٽkhv@ZBZ?:Mi4?d>;y>-h-?g7 LubV?Hl$qN݊8V@$"T'|Lȩ Ǝ'BKY*y5:˴/ܝ LP d%%yH2Ytd?t(@ VxrKd>[ҩ!@ HRmzg_ U&W#ŷG61*blR&P.HH?#-R)U-%]XMUFE_}٪bB~gPM0/2UC>^ -Я! /hKݠF_/&ܠ[O9c10=1b͓UP݌Mlr3>E)?WBK"luUZ'YĤĖ4s-%ΤCÙ%@ޗ-LH9x 0Z#9qTl5I%SZZKUKA!+ΤtScSM)*flUQ́bcPK6Ԟqfoƙy%˗˛9[(@63f_0ϢK|Uyzܡ9*;^6(yUL77AZU' b(/CXƳln)Ct/oOak%[G\yjGZ]h%fk4HZ:U y⟈2:S)| ejنLUDJ {EdktNR2p\q`1x 1;673͜5nE[ ͒ԒDu{efzOhviK~/>r&eW C$=2D2Fo\9:f5GdbذJ! |"Lr=@3o4̳E;&p^;H:TVO|;'svAP 2uX<=XYm:5%Η2\r&ɹɁצCNoq!ҴiD޵q$BSc&. `q$gsKE"^DIjI<|aOMwWW5U_Kyt>giԥ>=sm|PHh$g6urHuiSV @L%)~Z\\^ѕ߮nkEi)d?~{bq4~vc{zǯ~-g3 vꄟJl`FaVA#IyxKӃѰR̦vA-m,MEyÖZ~a#Af4I!mtrv9M/y L ͌vjoҴ/֑FgVn:\㻄ڳ^mCzzzg~U:&^W}¯;hl4?8tY 'm%0!/%I k-ݕ$igɳiu|Ƽonuї\itxNOOjXȕS7#.Ķ>QkWg󬝮:^]%ZA<_~eU3M~C1.V>5hD_ r=*Q%9' rVB:+<,XHUPuc@jZ~Qpޛ4qg48FcH|\-0'^4*z0#(c*Bֽiiˋ{Nߍw~Tעl؇:[/pvMGO|ͽ1C|5Kz*8r@lhu>LkP Iyk7AB`Frѕ܇bhŜ~ٹ¶m¡j[`P{c҄|LI5862Ī8X=oRfve_p?0$x속{>v~8}9N[%ڸ8vБUâŴLI/a0Tg5X9no >e7=($VIx'9d李dd.C'QCY ΕG{Qf}G=Qz:37>o&43/{i ƚP=2_9< Zv牠O[=6('k6* $Mahlќ CЩԁ..q9킷2 6CVAM 0E)cmTM lxBh7=8ܔg32.>l 3Rs ʉetv2@kxraPH +I9&M t ;孢s6`ƓӴғ~O1x,]o;#o+Ƭ|w2|ww%I&tO7)W:>xy>x<3c}@ X9Dc9M?A`*`]֠7$C˶]Aipuqqܣoѩw9 Fa3zn*jIrdWgY/ |5ySܳկn-|6ůaaaC{U8UI1(ΨL/''nsQ=*exk.ƣi=^xښo> p@ G$gz<#$ 4QÝ ^a]EE` "hҀ"l%}W|Kt#B %!ٔ*cj[Qøš+C߆U._YU<9U:*?}ꛧ-( ÷΅-_mlr~nGK%4, OLUǎfk":X*1 FJ]R2A'-̖dU˨)UO 2=mr dso0i SqEږ8{$T*laq-䬳.%*6,xCmd5$+w/d- l" (&F2Sl3d8+!IqrB# ] I0MM4$#\2|;FItY46_-qv[0Xvq$Vjwv+N`'$I"@d0ΆL ,GY*+]G8LȐyQ![E-Pب 8bR2HN5_ު8a/na<[k~jEe;Yĭ| s22' NE[ zdM1+%ji]ֆd@bgB jH1'-Ho8DbdjAv4ԁ⤙+1:͒ClEg;uӜAT)l'MڝsV͂d`b-q,'I8:h:vhaq=`–\^ۚGݮ;5sh(fT'J%Vf%c/;O#Cl dBl> {gDl}a`}7oOWMv)q'B$߃&B$%B$KDHTTJ"V 7RĐ4xd%F&rCF:ֺc2Ond#!jo{< E FQ-1ȴ1f.M) 1Y |s[Pg%ꚗʼdE}HU[N uu\[:|Wo]1+%ߛ,{j-edc*14SUvuwt+v f'z_ӳ尅sm8aC8~9:hkĦhx\Vr ЇdL'hGo<8cR!EJ̱\Ǡ٦ϋ3og仏ixxN){pmvPNY)h9!1"7b\O0t &%GPY Z3ȼ\(,匈7Vy_0pLlB8 1qۈAʌYmdgszI`PNw&hp1_^}JϼK(˫W;S׼GlJWʊ+?;^<}Ib~z9Vz?o4M~ g烔<}fNs\7?a{}ZۢOҥOW~Sǟw1ck]"h=|01[0IZ?G2/?oHȗ}$dɜ3:Z%ӥlp0Y]4.]麱l|R x/Uڥ'\ |6Yrku ϟ 3m7y͵'ԓosLB~gRzy &Os $ p n~tm2j\x RJ+';4ͱuKG[:*ݷ941+Ky~R|RYr22kf=P_2.;4N6yaqֻ9Z73I58UbV{P˼3h޺Ǥ& -Mb͋ mŭyR7F̺~zLȹHT3+uYeJ9bc!!\vH' J|-%D (î&0RA)k T he^pI .$CZ5!ܻIB릗 VOj{sK md̿WlhUfG?6/h&[T8t6N*ۃO_<+Hljy] Σu b4/lƣ?@/;LF {:Ȟ {:ȞplAt=dO:Ȟ {:Ȟ {:At=dOAt=dOAt=dOAt=dOAt=dOAt=dOAt=dOAt=dOAt=dOAt=dOAt=dOs=q0oauZUhG`h|_Yd8bﳐJ{uG}=zi,dҲ(KQ Fg\56&W*SD6J^8 `i5wFў&gHdӤTq9_ AJP`mLaZ+vYŵ9z|tG5Z~ڃ .$B`^)!qZSG[Zb4m"Z/F xhyO?MLK:u1F!u?ů<-NB LtpEO|7,E:<<.ЍDUzެKvYܮH[[VZQk$m+A$^UIZgj14 2ޭ&d<ῷZ#hbU%3ܗ{Ԃ5}@MPSt;-q}8bQ/J|d9;j%G-a$V12(:o㝱Vc&1!!ȈiDk45[!--cZ#gK*|F_)ȷ5VgqSw0`/%4F Ai 3i$ jxT~8, ƩR ˽)ײFfA 9|´o_J5>5[%6|_JT[<9YuǍ]nL=a{u-Jj͛b\XE5#RJ0łK+*x`b^A9AQN: .#XU iYFf}&0)ƃc]5"iY#N#vq;b$QaeD` -8!(\9!UFHHqbdĤgN x҄Ij$fXZֈY#~x`-IsbqɮzI;g\),^tER Y@L& oF3≆ӋЋ]!kYT؆!Fn~_qB~FH2)3_ "^)oކa,Q ,H}rFH/U2y4PJS=Qg[NPa$VٙRIa=%{ĺ2{N%䰳Swo٩;Qԝhi?;iTj:jDzedĈ1bSRG!E(ޮ.x7}|ݭ@F 6:$0s.ґQaY&\sWV{Jw[QY]eY[uaZ'pal?ֿo P D[]IIqt'[Ƣc\o{Rn0%sH[]FSĚX:&v3jC^[HC [CxNWԛ._Ge6K\(wpXzƌƁU| 9 NVFҺm˸T޴3 \_ޅb>RK6QZ|]Sḧ́5C9.O g^]WM}57G+7Xq+=@ߣ`@ U4!-w::?]&}4Cxd\ 3//ogjnGYzjXeweuf+|gݗ'[ gԨo`dI98יʧʮ*QOGUT8G3y50}S8/D0Fx|NnVo|ݨDB/A,MSAlj(P5 KLz29scSwzewBB2~ |YXtnкekpujmy>,#-q':[TW1OZT:MY6/h&[T9!I|{gEߙ5N hWmyJƋJ _>>TsC߂iFiZ2\X"EY-ӞN#=ӝhiTt`ךy"֞#FidJKB (!k}F9P2oe4si^f7t]w^zȵUŪuI~fjr d.VUUj ΤeպrCUY9 ln ﵞ(r=?G55w7'Z$n0.gn힣XK_]g;\KfLm-%;={n](BPq6W $x|8ύ3eom8pB)%#42҂BB19q1>p'].> ]|N:s(Ĵ2$:Hv40! /"Asѐjұio"bK(F[ :aFhNFc8"g^C,hzv~Ks]XWv'j~i3U*kVTLlTFLgޓ*۫%gYh5`Q1%ܖ6`U`A=V3xC6vo&ﱻ^w/d{1+Xgadn#[+ܺ`^KIVYI6 Ii(zSIP6Q0!S&aJ;fR@ \KX߭ C#6jJkl%61uSK %+aifc*.l~+քӭpHR=rG3wVyfbxd,13Qxo$#&"<@ٴ & |hZaK M$S1dNF08QiK0O tHdb *YF4?:8,a-۰ R?#XRL6.0Jf;d6cbHʎ!E7]@X!ݍβ ʺNFqC=v6wxrwI>~EbdЪ'?Ű]žT@u[G7]NXݲ~{n53AѴZ㞣sB"F 3-%+qp=v?YM|<Ġ߁WiNSWVDu@D㢶 -0)[R=zn<=V{;\&R(->H|NDpej0kd \UeY)\ІH  _ygLDiG0 i8k$H!5rݦLBZFϊLDZ|u 1@w0nY3>vifpnz"Kj#`Gk]Jb

6 LpO5Zks gJ*Sٍ@ \[Pyvʣg&FخqZ_q^7>D&fStlW;N]tpAy 15ZmMYPVg5?r)),2HApJ>@Zcq =e AJS ĩv!ă2LXF-Ghk1b2poF?.;Ѹbۛ ]ڴY5i?u{ vƘsNDrsPΒ`B"9T*FFC4R9XV[\.it*Ř[\T[xg ^kg=s`UڔhWJZS̊'m+gDbz Vc'i< xO TA4jt"FIǓdl_ DD"z7L#$tyE-rh7vsi{uڥID--ۗiYG<F;A;>U$ aR s^QF '] {{>d<^I|: ڿ+u nSuA[]'ra 짼keuvB V5@kLZDtpւ[Ŋ*P U(tt M)IObSGk9 Kd$ i"ɼo"DhX8I&XsP#C}EbԚq8  Zn Xx)qv[_s/ tzXB$!)KP;3ާZZ:FN5R2 e[5AJ9B4&i `Pi-=1h "5Mb- WSĎ9͔1uK=mQd-:1$UQQ1es3ą$⣎\8#kBRIVZM3G&@;56s F4ܒz^ "$rL7_z׃YKӱ}ur/zY9¯jhϧ>}G45ؔaNx:K}@Bh#ϛ DZ Ӛ7`ku8` uh;0J{{㝕K酦AF&mHx#F)jI<'[s/^B,vx39nͶV{}z|PgrOD|~&Ә|#f+m-B /(Z}}TɫTyj:'YON8*B Hc8i ۟377?c#--.RUL:oi@Rȧd-fw]{q qC 4/&(řP 8hDUh6I9K@8>&R{8FU{ݼPWi6>{Xń}S<^~;aX xgr2=kd* t i RlT)$-x3f_w7~ɎS +py͊Ųƾ!' ajKMbI,5&$ ͯI,5R5&$RZMbI,/!(%ղ&sTCM5&sd2L +ҳAYYYYY1=Z˷V`|k-Z˷|k-Z˷|k-Z˷z,5RZ˷zo[kZo[kZo{. *N6 Ѱfq ?VZVQ~謴YJN*+ dkU$܆][^ɽ9{*QVj #&Z)j*A4 .6C %e ǘaؘTRݔ)_1hY޼vS@,6>6=-Hmr3QdRL# NKɡpN(ăcXyCF؋FsfԵ*1)y!2q mĚD 9C<jfh7`DkR=b.gʿ ِYf-y{?l.fDiI> rMax[*(we/p("iUrd{%`6 Pj^wtmocU&i'm-?4l)qY%5s 2,)6\H$-)8h3pJsit>4stڂE{ RUPC @ųrU*?-M>rC.rN  PIkUi(<}L !>Dmt.57EɭyO+Jg#}cK!ƒUGL\wQlDZq ileABϒ'd<Ȟ\GIj+BE`F`+ ͞phTd)M, 1DZ$-|kSAҊhC+znc.$4tQWhc"1M0[ )Pͅᅓ1Y-&-sC=exS;k =(/)Nj9JQ 14Zf.DtP#H\(Z*z#8\$(ZNѾB-&?[¸S8.K-6O륚^kMǗ&U-?'8'IVw˝lmO=O Ϗ%˹V}PrOeBM.E &ĝSNPN5N k]EjBH-,nh?qPD -'ǂAV D tPI֌٭*taqO]إ;V]x,p!QE79gsn'Y? ο 3ؒm0ٻ8c و"@$/I ;N>8'EܥwIKrU̞ܛܥ8DsuuwUaTFTSuQ( 0H2Y\46)J *võ pg#9YV\*/IvIac"^5vkخ S&쳴Gw zACBh`ѐ( PD"XahЖCtx$-]BFdHhϣB]"ED: /X*FuTe^VFEVGc]5"kY#N#vq{SbteI):@:JzBh`̴HhG,e=*hI3P* 9-kY?8pt~|qɮzY;g"8.At#Jg P9j*eH 3xQuzqzha[ܱ>!EՁ^,sIi usQ!x1MP5NR Q!雏-LriӶmv(ږ]/{!jBDK!^(͹kTV%J3BI>]z" #clW`vz.~QO=ʨ%O$& 3ށI2*lA;1jӅﵖF]u*Au0tV(vKgvepӺ:a:R4uQjaʋp%/Z0`EψB*S:LQTWuKGn2J[\X\k#>I]VmK%(at\` 54rD"tJJ`Oa)6Ql_qq*u?>;|./9/gFiQ̾\>|w~_6Ի?Ld%\q5[sɝۧפ'O)ö>ֿO3^Bf_og[d#,=kVYV mwPPPOPI2.zxO1@CA3ō9 'l~ ^xp(.MQS{'M y?#r|Q+~Q&c<'oϔI‘Je)YN͐3YNj^P^0//NFxxo6΢kk=aj"dbu(c_gśRyϚ;W*whFwZ#;wްVg@wE߹Tjgjjr|ziX|&YP/俻O=7B=甹1^h^P /_Pj$zu ,vc^hLw=DK{XL5AK4CD4u(T%]SZ1m'9kX3oiV̛72*; 9!r%iuL]ykݶ>"ϧȪD~%_:֯Ϯ_o )R'J3^%)XHڀF+JJD]@'9 ⶕNJq/b1iV}>C &q)!1|nոxm1!*5NBO^h\a᧞Z6ʕ@Y[ JN5A FmӛvԮ N'J,͹ 8Un91q/"(9^,NÄؼZn).2N&sʓO>޴ۦ~9I?}UR7h bq=6!&{}WHDGG0޽bݩ-Ȑ=P]/7qCHeenCSGx73Q9_>rT(~̧uK?bv*IT//zu묿:"/_ODFoTE+Jr+y]O~?8y}oLA*"s3)|#?J"q[HYz+~>;{\*Z]Vhf4^]_b`}LH!-АQ:|zM,ePf7>!^Qe툳S֭\]UoeUK٫WӗJVYoUf Je#HM }HC>V'C/hنFj#g_0_Cp뉾x`h4ln;oaddY]:tv?wZ-SW-3-c]Z(u5wصn:7޹Ap<7&޿C{J\樮7OW8o4f}E 4/r B> "&0Z|jY"҉&Д:@ؓI>g<&cjqˉN&*J˽o҉ LuP4Ӣ5  4ᢀygLDeH |i/ >cJ|qсБVpi/z8cTPI\Ļc0!`䜻B*s^rBcP232g~={h]ƭR4.a Oc0z}=8|S zvH0HK i5BHࡼ^2A ήL,8 QJ5"Vt]-:x^^z,|~f8jRAA*hk/B ɇ퇫_E )PJᲗ&J1 `rv\4k D}RZN%(N'L \ U lR!H|Jk#?x\8scc ITL*$\gSG'LIExjt^kD ]hdo3]{?o7."DACAALf!fq:eS6s4rKR&JRVP:OEtH cnO8#ǙpN#'-LjFR3Jr2xL{y)j)LTG ~S8$rgH]c\k\ "8O` a<3[#g#t^>"tIwL2q?sqSWǗ.4gxFoO Uv T5,*?J|`şM%LϥRKڎ\iVè4GVPWЩ],y6*ṨLǮ2SW/S])2jRk*Vbr^mU.icx嚣*E-b{c DsWGno5j?SP-;;M ;L|kզۀyӛ|хT;-)pu嶅wu6Ft MQMګR>%dw0=M>і(0.|J}HMEd!],"ؤH.ScQK|I:g͋Eէ (хk§tp}E \*3݋KutAڳRȮB@&Qw_KͲ#oG0L3R/3~ E US`5.0`-OXUAEE'{'0O4v?vNi2Qh?k(QLnB!VˮRWHme±&6V:b"uV6 Ņ5 -cwk.4˺ F@r`^X4o:dZ{"XT!ThUkGU((|hNaRO)\b ~N`KU  * +5+0Ȍ̆veBlOs42 v?V^X szeY r3Jր1u2 Da(;/9 qN0tAC\)L1R"%8TR 3k'T bG9(y3XGfұK!\+B4ަ2ݙPHq Ls$eIj ڳFxwD"=dH_($_m (ĩ(Hv\F*^IZ#YUDI)be+%:32-([jkI,nJ9&dY\D d`niXEjE},B\FФi!:38oaP.irDw3RTiŬ䤡bLbb󢐴t8!bE!vvvיPoGQՂOOӮz|ou/c5$MG1*4MJ)h+JUe ,äc*Jrk=/3+XqAyjN$\Pd"U̫ fMʴL1hx L}辬$kIu u<o 71JrA#+PZDbYьmC5YK1Z1ةi*B?%n9~hO]@,M$C.2The1h6A@F4= Rnb:EjC$*zC݅Zc`8Bk ) 1E#`JY%PKcEX A%DEWHPl5CN5$BjF[v#hXA( =#`-yÖ Vn;"($'Mm2X4kn3"Ϻ|AW F-LʴCE5FHmr3u#&zށ U @ +IUE :ƀ6mӺ`fa@ZXf=ؤ=Cɗ U=e2 b2 D hs@:'/{idYkO^(Tt %׆.I@ b~pQi8 5fSMhpUf\. b!C1 PVR"T$4us't\- 5uvW4"坩y0B.8!# R Gۣ^tO}}7[˧mm FV=`2Lz8}}rJ4ٍ[*YolwF?3yfyPJYg"тq\c"D(`'1: j N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@:^'A]2'3yN t@Gx;2?xmi7lx~)t{C9)!&e4 m.OkQ"%eӟfޥ럶߿:XW{;Delw{7 oZ\ܥ?rQс樬EwR[!W.~q]ܵ#q>'.]~w;N/?~[zKtvN9^AHjȵ^j3uߺq͇^=6AY9͑mlsd#6G9͑mlsd#6G9͑mlsd#6G9͑mlsd#6G9͑mlsd#6G9͑mlsd#ܖds ,..mt2mx@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; NuA[{"f9\'"PFNctiCv@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b'; zɣ^}u悎RmpOvo._`$]e=*dzvڷg4(o0b׫ zVoU'5[&khwQJwM6ٚϧmo޽W[z?s\c;MT|ĭB-D|7>=g1g4Oŗ{`ʨ l1~lJ?_k$?lݺ>nSGzʸgWaeV¾JX]7/(/QܾR*Dxמ" Mom7bVWc|R8t)6kM\%?V6)Yb}u}V^ݍO<\)mt(gy?yI_]_&_kT免1ooίo%/WCM ?v1h`RtCLNe:ydgdQI2d 3Fy,z7^g:YLHfh*$/P V86cqKk#J^4v{m +RtETLc+gK+b~t" T ZEt:}KAW;٥>ξoZPסP;o+tܡGg(lZ ]\"Pnd:ZDW+BP:tute`/R+^F}^Mz14MpY MZ}4M([I>$ Rq6uUv:y9cd8~3Cj(DZW$j@w?ѩ稦)C\~ܑvޗ/,QߥWgHC]zble(;˸ef/kfӹ sIV>v8Ɣ@0cd v"WS@-*/<6/b? 0B)9D.ww9+TTu9f*zS%4˗S1r&Z+BW_Nź˩g[Obj`-OTrҧb%juԒwKT6vՕq"?f܋`!:>l?r{EOd]G%=pV:t1M'Jr1Ƨ\JT*ݩg刜ud{B*QбD%ճTWѻ,J_e4gB6|MV>70+t_Tgn&}^46]K:qJr!t>V7k0)O_ڑ:.7ޕ>"Z7s|2#DmP9  OAgUMM2 \QN5Y,$s/MbQT«!Th_Wp˛6޸geWv~XTsGK AqhCyJ?Nfa5hVoޔ.޽EMnGs'ZkQo%Tdbd#J Q|0r!XNZM܂%?f:k܈Ԑ)Bȍ+vr࿊ Ca"{&CD:X{))%Onbi1g;Ja8LLOLN4/|$~h,:ր0Y|e2X~TT ]92eʯΆCCn_a TV:;O.Ӄfz< X#S'j+39GWruY(j]]—R]&W}s|ˡ,yA>ŗ4٤_5_`OP5ݛW,$zQZ5šJFs {$  g[y'`/F$@M6wE6X ZR|y"6:tS¾FWCC3L6r@$Eә zQhleg r`R7se^E/{=*Pt:.@tpOKFϡLn4~Q*fm1xLH Xe ]aS@%`\?wL =q|oo.5rPMo=}Q'y¡c9S(g&\?tA&y :IvFmЧot=刭BqOEKOYCׂ:+1D Ra"{gΓ>d6T[IVGjzT9F:,D2""&(e$Z Xy$RL8ݲ9pXr^ HXǓUMn͡H]5BK]:*Ӱn!@TqkQ08I/ 2S u$ ݾqhAh4BkM̨3hQ9b0XJBԀPEL9%xh4jοywUI^^a` $d-)Jn dW2[D?fršZ[&WT-F+0{kO~\G~[š̮M['Դ`Mb^ Z 6ke-e`A}i& :k>`2$'GAv4Ҕ|;yC?+>ߓ`&y]y{_H0Ry/g\{!x5 1;#̎ Z֣ϧ`Ӭ`i0Btbnk9r`Q mD00/ nw/E !I%iA#h*e;4B;NxM𷉹^x[l`[Ѡ hÁ]qpO-k# % #`.\w\KF*v;8]z_؃hΚ6m vQ>VV(:m^78 OVDId'PBTaAΩқLUqX{D )vh2Vtk<#%1e(,:Rce"CJ!/ ɏ_v͓u?4O7]*8ΟXs[AXTy$4#MIqEa e`1Diq&:oMebja8$chQ{Em#N L-Km($uUaWErF2aج ^<83r%dɜ3̕tNNm 3X&oan;{MgLMʲa\x2A8ݗ!xܸz`6O>+%7_3P3‡Km8ϝqQ]}Fzr"D@_ORw$1+B֟nyt4{3(wu/Xr3Ceqd-]\!"bKnm~J6l%.+XdC7|o#k|cQ-|.ѥ;]sv_ֺ;MLq} 풀_{m]a΄)a$e(-y5 mznFkbbKesxE".$a?< |ե"`nKeP&uiU)*6QEYEBs9,؛wqi`DD&Ρ8{x .t'i)K2s&Y _3+uYeJf@ƥS1cHy=Ȝ< Z ^@<|<0B`J3 ރS"Ga('"=ˑ3(ewr;J/q"L-h}]xci݌{t9>(#Eu _u (I]M3% 9Mx-Eӕ1~lk^e)!E6310<ӊ#J17>T0ꢲWn|lAGȋIh!j\PB1aQ93{LRjLw3?(A^_~7_s}8bQ J|d9;ҟ[TZ5e&#QHYu2 ;cƌLb ˒豉֤ ihoc&7c[/)@ pzgqSv0`bNj(#Pq0DGthWPCR @vGE3Qa!8UaaR"Z.äLT:s$S&x_ :oOEet~ zH*)e2!%-@K9Cَ>$jO8͟^g5-9ee\=.QIgG,%%-Q9gZ1o!1%3,GQXPx \<<!TC8 a[vܺxٹQN;8^'^PQ"SԴkԥ5%is/jZ) h_4Tn 6j2b!4Qo;'ӌøbp9A_H~BAo },ŀHiICkS4NgЦ8.E1JKK8RA׺ֺ9#'m!jܯbuuiB57$UVT`7\tWW["U}MՓר2{ҕ~<ٷ۫,L~)~ezNѥhfRaM5g͟]%!NMkو5?9#tkilsdDqD&1N:t{V#f9FdK׸tr}gj-!kL$s{M9JD;I,du r!% KeD6W X[cL{MB+d_TM9!t@E3 \'Miz9nݩ䣧'()f47dfpMַ@Dunnαc)LϜgڳ+"Ϩ? !:2AsAjT@)ZSTY0d T7+EvǗ%-q9mVe謏(NYvVZlS\>J>|8#KY=`Yz~h,dS5ߗx߻Wv3U]%0 :#1=_=_WgLJ6]^gV'bN5K݀GMI˓">asx3;P3/֕(]1sN1gaT*3iA2ͳҿk3$Ɋ-I#,)$ZsE :76w螔H,8 5uIeY8;zlW\Q$ -)}Xw>z7_|[6Rq{Ͱ<=Ң3N!~lEa|0$ٗB$Fh#4dLedg͹A˲7IV Р'pV MΙekSO]RRR%. 7PwNٕ*!4RM[z?O>B4ҴDR;du ^(!\i1ji)NiLOIix)ER8@R(P"𹔌MYL2=$*#,8 񴗦_?ƣWWIK 8nfImfզ4䟯]HqЌ`znN><>4H@*;mLj{f#N;$.T2+[XVU}P^]-95*2ĒMy1h-,4oE1R>I¤ڏ~{Lm>Ҙw+-~.<[DuTc-W] xؽ}%⌴<2yu.]U[ri1wM@ Zyu5wR;7zD]L*?'H!uzhVԈDͬR:VG)l\.}@*SH.VAv]B`^ ֨7P. gW=`i]!\_/=:Ӳ )Qa(5:&$,6V#40Y'󍐮/\.eUhHOzZVm{Ͱm(k}0q g6R4Z~xfmXqQJchx{N)qsKGi㐰jޔYwtƎOfݝ^Kyidf m~0FyP`9Ŭo*ΨǷaN`&q f6Y*Ht\HH&Q<ȅ肴.E}i3p!A=Ƙ~V,*gԭ& B#)uK t'h/vxݤ.Q;f2$Ժ#_bb΄ " @]%"-uR!5:- Ҋt}3yǥBV02,x3;PR:χ$zҘ9L֘3x0* YZc_D֙LvYdEYJbJ}*k lq (܄vx|9}=hIӏ/_msuu@EFbtKd5c)$\2h)/:}Єㄨd°d_ wmm)9}079'" vèY)!)J~ÛěDɣtwUW]6Lb"J YFS#p[v&<_Zǟ'Am,~Z%K Ѡ;-fͬ,[ QEa` Hmb:Ah:fj4A){vgЈvkiF\":Wc *(O(U  'XMPiDjYF䭽UVZEA*HZI.'ĽXՙLB8G]yoe? TsW mgva(vgKE .{L4mbv?;Cy/&&,9E?`tG?ׂ߈-ŭ69eY|wKVr -\QK>U>iv`/sDZq} Cr.ׂfw#WE8V鼩S&~~q0zfco;2^MAaW30^|yOL77pQF<<ÌףTܻ 7ggoq*7+ߢⲙWqf0#1]4ESEruGBocic.ۮϻ,qvrt,?oE՞Uu.ҟ-9KET r|?{vLыweح9Ґyq\bQ*~\y}y6 N&|}:YJKnx` y8} y3!38ϳup8;,!6saʳ3?έ]YZ(OuGy^ku!~oP,4x2?CŨ&MQq%|C+h+ѓ~MlrqF#6!Kk$:QDym#.xN2tS5*)Y!y5oּ2\"44h&(B1-wdJ*j $E.b T|.N)!&_iWc\9fY6X8ljLM7Vy G[TޱigeRZ}k@ I *W3%LhehaO"ଵ:03AQ& )A]95Qi Jv#6Q-W>c"lYj֜\) qT;*E oZ jJ!\U~=EW 穏o6oEf>)Rdv*w6i8SD='T锂PN2o#[koĻSm~VYbW,k<1YhTXieB4 w`7^0) RBnDEJˮ푚@'qچnc-:r+j>_=gqχǚ]^v7l^7B>i%0{g1j8Ę D:!s&%MtDDI.61'Z29*VEK(D.D+=E̤TkƤW'07Fn^g=h0Mi=ZjKm݀)!l h acS+|\yj1g>&%H1 9 ;P n'U[4y StHnYK@&뙣NIeB[;%EPNT{{4+|5OԚ(qscZbY\[=հXiaVJ_}`sd  JAVNp"Lp>'-.m<剕*έG<n$N}yoJ@$w2HLtn_vJ}mZ{cc]aGg S*dևeݯ*f)u?b® a 9>>9=TZ ye-'%2qy~Q䢺R ñ4ˣ6 B\%ٲ (.cac[\~ \^6SJ=}ό?z]]&~s;_G `mXH.w6*lI+Md)Rr(7!JB$±@@!D8DkzDN+Zp]ɣ 9Cu47MDMYj慤QFtRk;w=Gxa.5yUsQxn}oEv` Ufg%g1[@”PeM/ 2qW^&3MC|l=yjGRb4ĄU|&$^DaHCsdn%OSzyݿVUk\- ,9=;D2 2z9pČ[)Tq f]U-6GCVvu+idu ĆyoAdCRB<3pfC5qUQke(8ȓA<ڣ1*i&)BȨd:QӮj<]gg\J~w|Yn>LWf燋r%qͨ}O?/'t~55^ {gvKȏp]'Vgmʌ/{q|m_g_K]K:_ӿ]RQL>̮}B3!aٵow{"1+|eFG7p+2X?pҌ`>P_+CDSQ-;3-=$${yyMf-: 9/5/AL :t^;zz1eorY6n2dNHZ:\C'Yk'6%3&f,]j^ dzj65]k^/nr9X@@0RRe,]v: P| NB?`2$=\ENwSFy_3'"L?q?ȗ'_]w\r;IqO!~wc ոqCe7=#-@\{n!nL鲃N[Nx3ْ !ȁP RP2OC?6&MFw_×H?S:!+ um-˯8ΙpF6+.Bz6% cg<5 ]BQJ}w}9ޛzInY\[^`;:6a?{̻xL̸tQv g@|htXZs/ҭP369< Z^_9~c Ps4~F%|6\$D"ǒTpEju/݌+Ч8N .o#tջb&{[n*] pʡU O^V=f[6h-1D 4zzBOh,h zZq!J 'Rk5R<$|IMC 4QGdC4F_q:dPsbtՆ L bT21C3@b,"RrxN@!9rS1հ9{n٧d)k{OׁLrD{DW \XJQ 4 e(P"(d}p,i(2 &1*'Ӛ06FnFGXq)PR?N6NEazhEdt5t78ly⚜?{Fap3H2a3v`wfL|%$O~UeDzrjF=8ԏ/YMZL{ɭljz4 C>]έRyPhOTSEU`.1'5Ex8Mj[zAKVֶE[BH3J|Ir艕28r"1ؔ41hqL:Jؚ!6ViYƞ,gaV,\)ͮid5z6y{i5nb|:Ɖ9p (&)qIS,g]J{$M۵ +`{!aSjSi.C)Itbb-59ۉǣbvg%ShFa s?j Bة+#*q">/3/,UjIir2 9]h5!J(lqJ*t)˖yؚ<* cgƱmǾD-DS6$4H.e FKx*KqnWJ*)%DGiGDq3>BQs#*YU [[>Dlj\.׏w]gkV/e\={.>hh3+Eɗ%>ޝRRͼ~ҢߘƱ@Cs)eZ!'@ؖMQ^3r,3^8_FQ/va\PѡIwؽ U/P8[Ɠ@-=8 xIr:-~rUq%^'_%Xixro?D 'Qr5gߟ~Qj bkXfxJ7(* )L2Kz8pKj=}''/:~@Ð?EġeD8畼jABj6,CQuZT.ˆZ~[9̮'x>IҢH5|X~9,/M8WX 5pck/!X".wI/j`H(7x>-,V)- Dis$h<,S_ݻx{LE ڥ=hX?ں>YR9ZԆ]F FVIUR-A( "锸$ǸV9kd nm>q(nCzbLP{=*fx )S8,gc%s(~6FGO (i/Fk^H>"Цb@[#Z9g3`ysH 2Wʤ<۹lDFQqހץ+ͧqwR&ͪ,[.x ~fnD>(m JmaN(|%sk'|\>s}\鄂S%O,?^(&EڼY\ b2w*IifޜaeFFA6&"Q%uLJzL!7p3dl W$Wg+Rk;?DWG+EF"&u<\Z;?@*ecĕX4B֛+E."vWxzgqe&f(XJ Hu Si{L+[ P,\Z;;Rٵ=W:MFB3#JrW:ΎT >F\' W$ \γu\ʵGlx'4,d> 8H.dju\J;Lj+g*#\`ϻ+T+ ӌ |U`Ͳ4ӤֱcUr{L#%hOF[O=$Zi-{LǓ6>ֽ4MbǸu ڦ>MSS5iYŠon'jԜb-@ߜ#mR}6˴G\l&JbZyTM1ؚp  H]ѿ`O]KS).[>ޛW/qSuuƚҏe/w3(`@ܨ|bn? @j=~j'\ ݣ>!/i;uvdr!r\] ~!lJw`J))rme(sh|^j^i-x)Н2[HH;j'I  k 5 G!~^ʅRT1ǓMޠn0̻RILI <(gk͵m _z6J7 mӔQ)dXeOT\ T_7B ha ]0ַV(x \VبLXm511 ˬ6 ,Fp&q//,x7ŪX M _^^pHɖRR ^Hmb/H*dǑXr$Ds/CIZagfLBh֌c̤*nDJ8TPh]|Y2m#$lh DZaCy0hy,FldFLhM"u0Q }¦B0yUĆ^dR&U.2kU1l~lLdsgF+DGHR 1s'/8(x[V\TJbISAX;VΗJ[U:JX<1SF(Sa'"UQ#XGIU]nIBۮ`n͠ F/AXżۆjbPb$Sa-F?^^^]@,M$C.2The1h6A@F4= R=ŨuԆHT—Pw}CP=n;H HjLEQXRr,a3m>%TB;뒄 Xut kR([|Pm!Q5MH#{Nw=o (DhQ:fƮ~6 3*I@d(J+1A2~ȃR!*8vGyPgUE תE¨2,,HcF8&gو.BPmʉ6BkB#`%ڳF& h%A J"-Ei*5.-zrDu 2QIX@iT_%e SLЍ![[&w<6K.>rܦ]͗vϹt"WcU P]!lJ''f= ]Z`-> DiVۆnMEѽT$I"e56v( =?62#bpPaRDluHmk"* F m=EףNJdt  J`Q Q%fdi SRC<֢7Q1l]` v+Bq&SK:)Hr;XQy/aèIP(EeQAR܌EEH,{ag=uc[sOҕȪTCтΨ19MAk.YJ@ZXf=ؤ=Cɗ U{&dd, 6#TK.dP?uN^Y%kP׮BT|ƛւ LW YW9ڢ>:jNàe+J 6'zʤ辐?Д n F{k|T(G=zMW:PCX6\b$*ێ׈4JC` Qc6ՔWn6˥!DL0r( ;fAjҪ\ >$CtO]:H_PFU~B:CwW4"坩y0B.8!# R GV//r}v-FҶ 7Yj 2 _|K`N!dzv[*y|\bVRn-Km7ityI;?n|?mҋ.^ ˋ16zeq1j+[VREh>k.o/Xѕ֗}leg+[%mOֳr6nںX6{-CS;0(#B[;%'Uz:N zȫ(Ўwg'):v@b'; N v@b'; N v@b'; N v@b'; N v@b'; N v@b';N dZvk=g(Jhr1{M[oD}{L /C~]/׷T q~R*F;/F`0}@PjB(ldPOm6cw@OWCc 1+Da[t3 ~>45{ - /{ˢ73VWWxuM N\qrѥ8_َ&94i 5Դ9]G[8(z?9z?\"쇶F5A-ןAqj8#N+c}b@ɖP 46(CF{pzi: ɨmES)$V*FÔU$t֮S+;]Je6tes rLW'HWFsmg+@[ކ@v.WtL /C~]/׷K w 6 V8+ vKTMg#$L ?zUI(wAOD3"+T hNWRq|tdaJE0ẗtEp} ]P`BylwD~ ]vZNX\=bк/ e;HWA9領QkTzqtE(Cd:A*Z&DWѓ+kTкoF(+tu]zBh-?I3LWOy{ꟽw]= 9(1]K#W+:LS+BݱP:%N* k/&CW7uLW'HWDaJӳduP۝A_BWFx;vj|Gp>ۿ Vܬ痛MViQf]-&^, % ̧r <{8:Ooz^VwwVKOrgza}}ui.ڟ.jyڄ0{kC٘^޽=a1>okT؟Ә7 {j`X ﯯ?'__~6mH-$V_>l~~uW3(m{j,==Е^gz+KCWM F:DoA%ۢ24q'Ly JYz] Z<S1^22npo:æ0~lBMd|T9Jf ۫Q9{ B暯kcT|I\,P(#՘}>(wO <^~@ǦФ)i]ʾbT -`' Bpʂ:y뷄R+^9hm1, ]\7<6-P*NUfBt圐M/}AtEh?v"2] ]y+b:{ VS+B{w eLW'HW  "1L6ptE(=IU< - /wή?>;w/?sy~ ˏ-o^ .wm_C \|?w@nvwMX,9zqÙhde{d)鸨pC#߯g|rP .C0 |]nx8sr=ASȋs i<}槿'VAVRV z~W0Ql0_TR63lc:e!b*P9DЈБxMf0N؋rm NWo_?_ tWrHx d #KkZFoXuqie&Ŵb)2B޽w`YP'A( RR8az/-RѼ8Cޯ&k <6%)nzak2C|X(=)/*YѵT1M{Eo>*QeۍVtRkn֫{fZA-ͳΦj{G\f3j@0qN# | ?Z??R!6FGτV()5έ!9xЧS"'47HъPPkuq] F(,O0P"LߦyV<͇õElJ| " zqԛAx %d/Y5Tr1?G#0M_GG6tO9l~7Z\3$M~ xŚYK<][ ,U`S7*5g۴ī ZHrv]JVEhX~6/ n9+2A^gnP3G+W fդGӆ{UGξݡ8jGFvX؃{;fl}x|+t;[[@Ufwg^0dK营,D{oͮsBH(2X|kYl8mJAk{6t"h>UTwqֻ%,J 蹠YH[&2tkDDY_πz06N.n U{ʌF g VqTs&iL[dZQk? WlO?Au[0_#u o'4U081UI^)ǢRQЎ'WA3F%Q'o3BDQ5rTnDGV]튉rl^\,j`snt0OǃHGO\m` 6.Zm{)˛Mg ee`Lʿ]yEgLz[Ueeе@h,RO@ ߧ֊: H uQMwlU,BJ)QeV$/^W!r=L-oey`H~{Vw+/@ bi.H]ryXHVl܁Olz~T6x'{=@*N?.TzF>#TRgUZغ͂B_T>sfj{K ʽ #o萙{uH\amq0X%* ^^~- s$˛*E6yS$?07.1餃II$yyZ8g<}*?gBʘNB81x O'O9S%f[Sg7?ЧY: )3/ /W7W^.Wb b LYa5Km =ճt\cfuF\DH$ޢ.m,[H&R!R@RsbBNSO'>->X"9.5%)@OQϔ21O<\J`J Ő訹wUZlDkՎDxDe"X\籖[iŠK-Mq0O#óWmgu=zMȫePeL6_Y+ܰcm0kY%^^I՞x Khlpm .:ZWJljiHaS;ᐋD#-3$S[rD_StOӁtOe3h͸~Mm3O[|q eĊ$fB`X2V{e{,%SǭL,gBS1Jq`\AqY p1(8Ra5 eٚ:{G֣ɀz\yKc&k2\J1Oް&x_,+WE/MohvN1T;&[Gm T!XũRp cIDC [S8D;OĊJF`s^% ! zeW{rI(XTFqYƞS؝uJp2oh)^[tL9Z'perLO^h izg1|\w^5ӑ Sxc'hu&JKAzLT։`8( Mx HvMo8OA 81Pdv11/"s$ uY1NXc|4|B Q֫]f4}[:i#߾+6wtEQ LCzqt諸,@ŗ?D5g*BLv>H$~Lg 2]S|& >F?xr'?W?[L՚#֘*A9 ?BZn%/sB@J4>x|ʩ`mZ+s|{[]8; y[6%(+*iqSl`8XGJ<ͳ񷪎|4qUK]o<߃J6墖[Y0E} $0>-F30꤄bpz1:bPL$l w`(oyt)K-r~+PlѪ /9Ge8 g~TnGUAe] Aտ{rQe8-.$U<9~"#ɟ糋w+%^73-&cؚb}j7dr<,s׿y3-HSZ~m~rp)/*Eb>uTzRZEaq/1+~h,_bxQzu?\YxV^WthNq҆3r[I[qYb\*ך a]{-b2πȄ-bNRR8AqG^zd0*a*zS{` {Ba(+ +]P@q%)"\F"s'vq6~n)7Q?d}'cjEdZ &`j0in2H.Z1wAl"&$ .Swu`s F8X]85u7 }%wS8y' ݏnNn8KOU{<϶6Ɓcf 0*R9ć8Cm]3/mO1:FF QlMSUk V ꋿr}  u<‚ fZ.p+롳֑҈# jP%s4\ۘ2A J4R(ךZ ;+]\BA:XjwGMi~a8j+ ձ =JaɑO{iqN{:O=>՝1r)|Z`~1 b| 2HL-:"7KliSNZ#IOwJJ<#1D˝'Sem">b455 F& <c&ˉmޚ:`LQ8}t᪟ ePzpw)i_t^~m; O CS;ɰT($༉<@J)pN뼭tފl~ JS蜯7f;mूVP|Mg:&c>|D#W7 NGuF@Ȑ#ncB 2Deݝݝ\ݝ\z'vIa$NIjQF'T/|Q{`}Zcq1I(GʁjL{H<&FΙXPq*44Gk δ[Sg]g=+ B}~=TJ<,vQQދu57Cq$@_?Qil@& DžsN$9UPGN9 vo!"9Ti%62nu4[mMr=)ZKfS%m=5uv8+鸅ۿ+ؼSW8o<$ >,/_boV◓ɭnI*K,!T4N8RQca* ծԻqsy z9^,op+[ھx̯VoG68p)JB2"'e(Xy >8YhÎ׏$_9JAi*`veml(̑( '\i<1U7!JlX{#bL\ D+opxH!QC@"ruUw2X2.@p3j]6Ok 쏤PKEaAH"KBhPX3$FbMd^͘m$Q!5L7LӫpM `΁12R (~C\*b(IQ%gDrCuD-씳)l+iSLu+c$c)cj~_=c _m͒/ jvbVg%G]vHח lF-Vu1#< z5gUF#(["RbJ^Ŵ(,INOkt+KI(.p@KRzL "+!d`$PO1&d潆Vtacp6SwYhGW>03 QwYۼ>?#P`|!emBFL&VRO<_HJ.3!R3!y 2beA!n:Џ_B;E4]@0s{D6D¡5]`+q\dJXHZoL{۪ }ф&PR~3@h KM8 p~pS§4iYk'ra?]JoѷvҦƵ_ybƫT2<C,|ݸNJJr~Ր>qnSEwSqq^wwMz *b+En5z wߏ~2QE\!wJ=7\ |Ը>nnzx,pٮ{{v|v*o׈tE{tZZ[dM7Sc}ځպ'^Nqn[j(`616AIm-z®AT[| 7>oݓN'ncHnabV#{Cg&>Lx41eZ=w拦sU;f, vd Ś{WT\iO00υqYex &^iRzK+_vA[ǜf=s_cNCS ?}2pڨ 4F ?2ۉs]$" X8KAgDn7Cz8C)Өڞ{ LƭeŘ{a@C}S|S/7=Wpy;PO2F]Ú8uʄlm 2Yq!# 7drIjmpmugUݦȗ{"_EueiLf32ۘ[k&fLV2rhblaeӨ«Sr))!jS^ (39Rü+jXm R>=?LA0~9%{4I9(_S&R)C1YAMh 0 6$nN`P|{ Ix&1*/ךа6gUS\Dxx/Vjcd[+6lnۿ?>[֟œU'Y!.7U=Ut9+C?FYyAΧ3I*eL Vj,*R<Ԉy`LF9=1" 'Ԡ9y]>*ЩC Hifl J44ƅ͌-  gFcԌWlsNnVɛWeՃ?iq^}匭(&&QpT`Uv `fYcs&Õu͆Bqóa@cEF3MtP+5Qg@a& rHm͌`b9KCcOX{ۀ򖵟k+)أ:h:HP" hgGnH<4ʇY*4,d +:kb"ALP8Gc*Qs acp6aeO,Fmj~lˈaF-#6ʒ\ye9yp+̙G ki4!Kɘm(-6NRpj3Z<F$<fBṡK6g3#y|:%"oYˋ-/0grX9KO"ފDXfgn@ڒE.xqoٱ-PlLJ; AQ@jܸx8qz?~!GM\X(k]?uQV(B35ig\G:`ˎHD[eG#Vv9BnyGHbepnq#LIUȝG9 >Fpw `6( #cۇ<ꠋ >{ %Qœρ(%`*El ~!xBVJM4Fgmƶ{؜h]4[/(@{=6XgQzvB+ffTKs0҇RUmgX Y` *-NWR薮!] b8-\m+ʝ<-[&+Mr~0Qy*) +NS[DMPS١C$~S)N\H?1i J4PhѪiDilKϐ v0tp9aBWP9]!JIZztd]!`.gP Jt(u]=GXACWV}+DiMKWϐFύ#,qZ ٞ3O>ժc' Wy*c'2C;Bqޱ%2Vɼ_h;;zZb̡<,}ZtqWNOBii4U}'8-2Oe',^_ GWo^_ZY'/ `)}&ŰȦ=|p~;9.=|'oːNgκQ}"O$k[0O ^mn%y.?/©fCyzaGŜ']B{X}xpiNj0_tbj"'1jYLq6"{HH8w{ib.4-~Uնȼ=c@KS*ΖK'׫Vnћ'hoehyZ?[15ʎf]#".n;css;`"e#qZ""yY ƒUO8tb7a:jJ-+Itb|f(n5$UpS '2*R4 Aš͠pr]$K RS@,Dc$"@h%CCw3jK(1L 1( 罥BDR> KYJ{ blyd|>5p,oZoS^ 8ͱXD圣Hqa)Ѭ1@h-l2 78HP6ɍTy ^kbBZL5f'jRA7PXj1=ٕ,u)'*Sn-wVYnY AO窽b=O%EQy cG;Tp+@l"k3u*758o%cRF'=1$S'Ԉ<$BLE%%4R-c1q6[rX-,;{g -\Hmvef!Ct9pCRࡨ=LReh$ȐR,:@$ALPF!! Hv=,&f{XE6cKǮE\{:Ĕ\ye@C  &sZRdPBȍڔׁ,e;#jMȤi*Als1Σ]ijya:ŵ:&8ŃѝRhEl81Ж$ ŋcKiǮPb7{PTsЫ#7>m'ظQΎ[8uޏh W2(VZXЈ EΘ9X2ǕϸJ@ptYy-"a{rd!wo;I'Gȝd)I Rm@wH8(6%Q΂.]ih.wN;瀼ܞt/z)1U<!ȠV)b`أP!)F[p5;W,s8E- 7nݿoօzqגl]sB\p6In̓fӭrԡzm&+>ySXx9zM_]&l^{<OL`M.1$~<`w_$|Mf%r ,"%o#jX㬞oGdY6P'@ #-e}>BͿۢ͢2y S9qm!<l.O5흟E 5kd4mo_;~C/!'(fek&6XR~ػ_Nn¨N.Oz-?vzطR|qu͗ڕH/D Bi6.U"U"[ym{9حݵ޶KroJNվRvؾqIA}'/A]TBKkb2c QAgiHD)!L*ɭ5'XIvl\c~گrpA+Qo2!!{Oc#׶oތFъ,]Wj"we^곎snNW/`9-#| VW]ޅJ/FUVOfn?wt35cSkS܎FPù 3W‘P 2̈8l4@X|cG3ݩg,L5g $^Ik0-It\!K ^kϣJ3ЎY1RMs(wA~XμyNng. y#ZњYe 9nΝb!5mǏϥt_ٚŭضծkj)\"Ƽ楒a2o~wsVm*ǣ=¥bj?^t}S#u n_QkqGPscY?f#񋙺c'BU] /}#5[sBʾ$cwo|u![τMVDR J%-OMO)kIS_Ar?_7a sH#c ȥ5d- $F<^2^UգC\;bRV=1`۠* BEIRA$A2M麘NGREbv:[vӁ{-vWz}ًX=/av{>B%u*)OgsNF!Re1DV>Wy%@9e1/9ʽdNwz#2_eϺ$ksܙq1R6!BL"kFnN ]6GKYc( Dx0H   ZP B2 T#JqWK'}jK|QcW8ѷ! wM>r7^qՓcQS#61 5zh$Nps+Bʑ]ʱ,W"-zn ~Sw7օWjļ) ¯b,Mm it#Z' 5SrPC Jg?1lTB !fas Z"UbU,Z*Ü ';upO\.),! XFۦ}SW'FVL|$ȵCpy!}DWY0LS>/sWtD䓮Tt:E=Ш,O1FdH*,N"R0%-7i)Ax^3St4ScD k?ߎ֠yAu9Hˁs5]ܖqዳ@]v1iֶb΋eM;{1Y?Z0wΓ2c l_>!P1x2[Os80I@"&:zuv2xƮ}b4~-K<YA+ooBQbgkٹ?+֢:gdaPӟFC1.\R5O'ņԽё4#\w[*.cp)70Œ80*ǣ8<;j N A(mcC')5]˨5H@eW[2 tRE%q:b1̜Fޅ/__H"nFWF<,v[%z] jNY"Q1y+LX/]J9u?a90Ay-f Z9L$KRW\eONuB5.Ip6 :>qm6'ǍB- JWlkq X yS8_5 #18E aGZ#L&DS>tƦ-a40I#}Y nzb!fyjW&y&)ġk#Qe,Ǵ:z ar֞}/Qo<\ru8Kkf9 .%]{ qDZ˘QxW\0Vr8tD4bx'9|=TjJ36S̮lr<h/ VKiN}/kԹ哵̊zVu/fq)d./HV'iWn7Ku͓7ʑsyA#~?Fɭ"Gij8=#z_YlEМj-4|4K.zM+nGt%Jin_fmn˜/7(hxT*j[;z I@p{lW\V{ޯ00~?nnlhOPo*ן_Nz 3kE) u.q9`jYZ[ -QxxȻfFޣ0[YV9$#5>&ٽZPޯw u5. {` E@UIQԣUҳJzVI*Y%=f{g`JzVY%=gUҳJzVI*yx04ʠ)yH9s\H4>qۨ$ B̾2d"ਖ਼Zg&$3u$G==e}wKRm #6{(o([&S?7z?x>̟*w/_/n4RĢ?+~{;Y;_׿m|x)^lAcF~ޛYOȶ—;,ә7<6"0cDhqYNuTZXpji-M? ..R/h6w+پEmf՞ޏ\HSy3o{ m JbZ&NojmJ*xGjIZtvɬ| Ԟ`˶/j֋6RÉUBJ͟Nw}H_4322x5iFZeGWI>~JFt{9fk[a߶g\S?O|cоyPNݼ #-`OihWƯ/.zoKy[Z6B+/.-x>m֭lTZգIlk}1+~_X@{4ϫ&Ȣև#k7j,/gj 6m#nd#_qkFP:XrBI܄%*&sf| ^L`Vӎo yF(1|4DqH9⠓b\%'X ; Ml4=n߻YXY`;.} X|%ur'Nll7x|{cѽB05'g V|D㊹lD㊵K]f0q2)alݿ([QmB38r'AF#xD{-uk޻dݟRɈ9lKu5U !DMD=̔4H$P}FӭRk9 |01x>WQ3e HwL:3gDs홗qz1l>`Ѝ-x˻G*v:?ɱ#9ťƃ]S폌vV>JZ7+iS4nCY u=h`̥Ydi-hki]\'ۊRRimL7%j]/t :_AA*j+kc㨱w3Q p)Vګѷ(I>)YHN6^ h΄XT;^j@YpjKμKj)&^i  rgR"e@NH ()reFs))By;3g7|^胦qDvǭA 2yo~ Mu+B!P6^gzS5ia !"&:zU )z;k?-Mj}_eI kw~Ů05MnAڅK##bxTgd?.v苒dLQ`Y.B0c2ʕT*rs"7;j N荊ǠBb}9tRs)~@T36 t!eɭ'Ιe4(A J j62ucؙ9;By"_N_݌>OxX춰&gjh536%%tC_c^QSb@0 =\uc^inor=9 9׈$ل*80}A؜7Ft4r`ܙ9GZh8͛-?;U59S^Q员0Ivo&zleaB*"K@t:z a%prஃNy#M ozc[E?hNJL@D'zCi%*牓JG,*AӜO,vя{3'ރOv4g3i:@[)̱(UQb0ICDj~6iRzb%ؚLC iHL_j4Qjgl􍣟 FmdU#WM+ '$ Uq&Lsn ,%W"Y94d<ҧyXrT)ݺc3HAC,z)YTYJD\+sYRa*|=M7S+u ȴ7T! eL&G8G"TJw_ԑr:#gZ82ЗKn=Rnscm}H #G1_u EI|IMfOuOU &F!ЄxāK< I`HD+fQt`B_q$\@uXER&K"2 )15 D4V<$[վ"nC 0 0M%SO(h! DŽ /^vIFs(;rY6,mj',vw/i?c;qi'ez}~ڂ KnӢ$JђdKk,)Q`#AXwHX%,Ѓ$,X:Ϭhbh^R`yL2&܀v1hZuP10hPJܦ@Ew9lVJHH!mynpڡ`wLKr8st.+raO^+G>xCG> [>?圅hp2j;' LPZjMQ/EP{a;w.i24doAc[Y ag$nj1dF0fM13qf;/J6yu*۴[%uiW¦O}ۡ4!>dZL]ńkV0֞!#Ào^#܋FH|^1!bmDAQdtDBuǵ$ql"L8Oez+p݋,1W;]/:t;n,K(eIxdO!DyB NY*,419Ó4HIj"a kL4VpA[#BpLdH)1!ʽiE$n\/ɵ,O4]*hNq=xRoaQ W3i^MYP& Hfp/`Ogb/IFQga - :nڶv=0-l /ē׍nv7WƝ6Hb҆d2F0'1keAq)R)Uho6nƗ[W ҋ4f4kbWe΍~6G֜Ո&#Xh?1)ŕk ùK /rD}~:qch:D)zqx1&MRxRrVFA睃*H܇hI>/%JLL=M,%s!X@'YD\,]K͒e\o>XjSӵ~KBa<70\-P}%=SbW}pibu&'3ԓ'fքԃ>5)r8u^o >wnP9 H`0v)n}>~)k2axМtG-a$V x$ )[xg՘ILRe4zl5).BZ"Z*{||;>IA .;p'=z*c P ɀQ e i 3`F QSx <8}\]`hT)JkYP[#gQx ~ b'Ze+u:b8.S%v[ <9;Y&.n垲=nX{?ܢU~2cCLG;6U{#BV XS1%M UKuJK(" ,(&zSNq oW12#X %PRmk٠|RkX5]gl 1' gM /qsV7n ㋱'jP^oOr-0bc\rD-ASa(3K+#/l.U42dOP& fh/#vH!хTmgn~ܝ}cGNZAI 0i^{X ԎB1"l9@gzPtԵ#|T3bVA9AQNpP:HcZև^fHqgpl[F$-kD׈{t 12LByIcep˩KP'\9!UFHHqbS:/IXobA% 9 H2^#s1Ƃ^7lKՋeHzqWrl;Շe}ȶӇz5r7Il?*";n~|GxJauxJɴ +ψ:@Ȩ F">eJ)Ne}{ $d!^!!@HRQ(먡镑#LjLSK0Pv=9uz$d2+osuzun&'n2*h!@!eEA07^ye7Z{^k+U)QW PVEcw`1O魶[#ٮ)ti 7:vA?+0_lDe2f-pZQetPzW|4_E%HRi& (KQ F%.u}͇ +4 0l4S6yDtyUśHK/BVn'aR+nJPլS lJpWoTI3Gᤛ2$lH}X"&4:ЅK?(Fw"{'0ta^w,{ջ?>D$ՁXq+A7d@杒ˆ2F[nQa^5z߇gYs79k|fFA+N\k Xs$p]"3VZPvԆ@Ϝv]3v$_ GŪת*'MԺZX:ejKgFюs;e65=ϵ\`EV.GFFg}P{6 F]EffM՝_ 8?ca/A5r $[n{qlsE 'k 깨5VT"@%xk kL$ea]R'<}~Y hYB )ܗ(\2ИqxRut9aZB`YVw>#'ŜSe1|}ؠPLs,X1ͳ\vvpձWU͗#3>}= _wK)Pr#,v;7nٻFn,W}ܼ_e22N ZR$; }K-_JeʖAlb"?\]cO ?FI^_ z%:wƺ*h΋K.! j<huP4 \kˇe &2P7?c+p{!zwVVd,xh%O8e@)#V0d-8!dy`9C݄Zs!QefI"nF)w,hae2׎&n[m:d뼥7xe @; `k-"c*,"$XDJHָS)}oy#>Shq/PLnjkT )Bc8>%lqO٪%=shXBqHD8 SPQF ^>E\ᔜ.S/dxηMqSx~ '1G^dDg &{Y .t#Sk '3yWdKp $ou7'LGڙ&OtUn# -v3>7<9 #9q!uo~*0)5dc )7VO6(>.9A퍌`™ds0 jQGEb},+B>`HzdD:>{6ݭ{B7#6|f>>)Y2Ds:l,Z@ I(mk *$Ϲ> .y >C].BG#MYn=ճ>XKrG;Іm seC731KwJK=ࣙ^i?{&u-U84/խ47IxwV[VSZUxj,u_ϟ&ufqHeNJL( zI'(j%58# '>O*@؈ޞ0(ߓ3ƥ݌Danļ8,ZKi|)@Ŵh<$d2l4SPS:Cur[Gg{ Ysn{nV$ݩm\r ,ų G:V'/c׵1("ҖMKAy3lriqshNy-8u9t<7ÙONZnE)KXL/+~Ė,FU9Dϵ&E J=:("OυThD@6ª J턶 ֔磂)u2 @!'=ޫ'㾫3tf g89/*AF"CK&3yef )tJ*vi뽘uXжn"6'lqzu1b̑9"㩄a#ngϫv`4%OIͱBz! s^u2\*Р`Z:nh)`Og 7!ʠ;9V~L=HhK*-Ơg^Y`. B0Y (D'82՗zz?m6y$K}oyu$JJ  ]Ve*p)" J˨L"K_tJlXpg;~:y"Q#Ȍ'N$07:t1x>dNֵⷧf<|2oF*/ݣlP'NfNW틿q&gi ~]&=H`)3{?oڽI{X={Owפ2m&YmY&'1rxQB/7#&H>ZF*'x VsGȩ OK1c^`!LfcE2Ol7;,smߕ~q]_Mc!BT>' Fh]FpbҌidAѸa="X@C Oߕ&yr@f@ mrҗ,]%uX޵"*̓uD`~͊ |Et7%+ M>QA{two-vPuȻ%*eWxi[-O HLL~Z$ PiqN..#i0X$SuDkrlCy4-ηù; N@u0AZZ0cB>x9/>RƯ]bڔ#%l0Zo淎6s&W c),ւ,S-/r qq1|_6\ZᝆrjV|„mrq_ vx~\5-pYAgg=Spk㔑V . DeJ_ܼʳ5yo5CIJ00!|oGY{[Zu6{~4ῗ`ML޴0Y;܇ˋ"W1?1(ɺ˷w_t·ފr{Mϛ]ʾ4~WD7IْS<" B|Ls&X VW*| B~w!vyNi'?wAtWۜO|B` }%У8Ņ,>t}EZi.I8vr[̂T#n[$ZXSbG7N`kݢ"ЪE"0n`4Yx@V;Ψ 7yW=uq&nvQ8|r5_]c)fPa;Zw^Q[i< Y d- 1+pi[oЃ{1/N̫H~y^E*9~w1}}`Ekf)"3瘤O $)Y Uv V\IA_K>C+#0w"]\~jc9m aF7C (V{uUTZ h'P[ ycrNCzfeN^ rx` %ڔBJFg}J%>Lq߬҄+X"-J^K&~ it*A*Y d |ʧ5)f=?BC3 t!'F$$)n@#X\gw@[g{8W!W~  . ]G̘& _̐"eQD# 4S]_uWW;gǷJƪ+_]mE;_vsވ]su/Ajwz|se 0i48{\y iJ) "T(wd!TY@T*Z+Iv^ Sgw弄y ;.ȃZ#cTȫ2;Tp@8i*3PVl='dLⓦt?"P N['%EFQ8?ݝu+A]6݅JrpNrQS$[z$KEjFY#'' {M BEI{7Ihtmn۬bny9,,Z7, t.dT?};l:?i\1.^AH/Y|OgĶd]cpP^ۇ[َv@8\LN'H/O-'u= E:9MH@Jg+PTT˴~@nq!xbwtGCL2pV IrB4uG5yFaY0j-Z=,puImC*Hos: JޅT Yk#dtWlw[|L 5{UWg=n15Df[oFg݅gn̕1{@,% Uqz2Ve.MA* B/GMxA:u1C=;(x ^l/:3wf͗JZg&!Hfx:W)mRhȋ+NQ&!Je2{D.K `]-=)`k!1ƱuJR \ꕽv^Vk %XIKdž=bo~-nInJA Ӳ֣ r1YP"1O d7 <חS0<̲BRmgWx.Tqt1+5?mj4ʱ_ Qd`0M63ܐDY5 zQX3TA8fUkO'q +ha݅-Kf ܒM+c1h&l T&Oi~a8LN5<:34RKОCIGƹ6>de8FMEfK^}6R#P$TR %<ࠈ8:O_qsC%JA+ob~gƓW} }},cRVj׿5?%rE7*܈œ#5S_情[ rtqell4l6p\/5ۖ(r8Γ[*o9?2jeUސVȟ&QikUm4o~8e˒r5dr|||G3JOT/N^BE;>{٬dzbc/;SPE_- |瓲Lg67M"q՟ _Vv NS܎R{{t9/NN^jg6$-,b`K=8bT-1A{(^ q{cX/$5Jj|m$*ڑf/X^2+[\Y6''WWEN|46 ^n"f ˔$qwY+C}|2`I՗~?.H8u{n&QLԜc]g ҇Wїl\ImL''-rŘ{!k<9Y>3uKM4ciÝkv>myȁF(V+ Mzd ϖ&ըODŸQNlTӶfǍm4gḁ%$d_;X9 !$߄T*V t46 Rdh`v+ qZzuk~G^D.H1XɐDFŸB/fx ^En6f88?^ tZ{^n:0.0k]ˆnU},9?{; JPu+]d֑+ȕVLU䢏KfuWԽu+If,z+֮1&M7̔4PCi#ɜMjKx}M!}&1#.s9 %q>YVϰizJ[S,n5_FأAGkXn֋ _/|y ɠi@df'-i.]u=kuIiIٝ"|\#GURʒ iz)W9sﳪ`H˧}ϬgVܵ-^h-1>bLZy#b$-w mT=I%>pjv ,˕'`ReUGeX_hղG CQ2 u0 <0oUv UK`MtIDցGy!eg,)lX]ڠ)O#Apa$p:}|I3=Y{w189zv+8iM]D/%zo!m`&ˡNIs`4Y%c^F1[N5&C`Vy2$z0ZSc974ׄ"GTך3rkJg3SrG */ 2~z3I<.gZy&Gv[s6e\#)M:T%]NQIprB+ ma+30g'dB66QCАpd"&9bfخ tEv&b;DkV_ `'$I"@d0Ζ:e9nC}yȄ 9%5 Y#ċ UErx߱>쌜v'p:lخc_(:ֈ׈F#Ü Edd [&dR6fp趜1m|cT p \LH!Z@6F (BP[gl׈_#Mu 8Z:θd_(;֋׋^g+UJ2ItU`F-zK~c&%` D.^/B/>;Շб>Tز(҆uΔFqa&!|MV!XVg=b$CL;i\A3蘵'lpɻRөNRǔ`frpB I$#@(E} qgl%je3"+ӍrhK*;b|E.Od k>Kζ/=R|zϧx iT%ayZ9*Tb S{wP?Niѡs?~Tcн_)-1+ 0Θޓ B8ACQ:{/c4 | Q;4~}N~N0eB9eQAHo`*JZ aF8F(ܲqzAM(B,bt5".3Cp,e)gDŽJ_doR8& H3 K*2fݪSw7Vjq{7jnվ౭]޳1&@K'EpY˵hO%d"+ 3L^tx&ԃnTI|\Gu^&dBeH+ *_;?q"[TmA ïsLFx92ab3>$uݟ1jd7i9?~ToVbjVYyEG f1l0?&z믜CnPنZ\ղ2[Jo)-ݞ҈@XW(UȞGXO/I<\O~\+B>E1%`f%Z<هSn].{Zm㑻["Y,X󿺡agzߏcA㿻ýxrqv5[1e_ wu%Av6✭zGl&&缹ԼB|iWDzl &7T CNrTߗK"ڼ"vj4_RU;Xoz5>LƳq˛HYSѡtiXju}ܬ;;Y pDvE6S)6ӌtT}|?]R*h0NK*SDv{!!d :H@Lue $puoo'7m~ӇuproM_l[Xoο~V6'JʎH+#ߨ Wd3 iySYs;?$| ys/89~j⦗_'4;j8Es4]y"Ed1}d;fP$4u׺:>6ist7FlɛLRi0d5Og=J3۴XfS+}񢹺߼jѼӮ)fPhYwj&=`\2Djyj!7Nr5UEEtUԎ}zb^Eytւ{@vqzQYII]yA6g ǯ%fuHݎ+h7Z˴\<=Ce+`#ziOꢩ_&ZPbhEIh:*HHK.╺h!Z@I( jޒeɌ˽¬7Y?%*CF1[){#%l.;z -IAVPEp o#Ąk:gDsYD2%FB! -돗iyn> ef\"˱}L`z}[h|7ڂ 훎rNWXm1I%mߓ!ޯbBfO(q+f-^AwcF0:ԫ5ɯh[p.H P' Թ`a?isrrt}-ѬTnKm̱*6*jB6*>tLyCva8i*_ɝl"6i%~5NB"R*P&7QKΞd8A#Z p'dy)be^BG:Xu8x7|e8jVAT0m^Qs$+jr-}-ϑZũ?/TJ??JZhzVdzx2]]*[3DI_&:P4'~up' bQV0Yd&7DP>*LSjNHɸa]ʐQi`YcL{Vv@okC M$\6⼰KdJʖܮΕ\9k!4҈5ղ}wJ_ ɗmGs!)OQYgRit&7T3BY]yؑphigBPFq%Ǥb&E`A$P ))>L2|jgŦM0`ͪ<2:#8j e9ª53{;ݕ񹡪6;T:Y9 U|E ORS'fLNXJeA:ٗ!;*t ݓhٮ_4g`S[ԙY32* fYXckmqhZȌlRTPG̥8QƁ A'DJi|Hn4B(5r]Hӓ $Sy`dn)Qx|l1ro;"ʭMϓn-n:&Ƀ,B "x&j@H.JUNw h4dPn_4=n^ d^=޶=T\ĵ"Іk:kPYF3qEMR(+phPNǮNJ MT|bB\!op+) 836)i NtAJh_Jnkz|B;(du + eLiePdh5E&5CR ))1p#QJK!Q2Dg:\Ó@降H \{Lӏz/z' B V^>7 {x_hN3b`Z.Z3Ib* ^y?s1?m0}$siTjϸm.-S4/"}#HuAV">UybId)bWA4{Z],$N0b7[oꌮrͿ 2ԢZ?k} OƓ7Nov7Հ2l\tzV25+,O:7cTf3f/\.EEW/kR4C&k"*vw&vZ Uvv9/gZ_B.Yy]5]he_s4C ,+l8}z*}3ET/Ti=i-}e~^eVSIUOަʙ8ɿ. z.&R[Vb~[y1ÃBKyXݜWVgӧb}uŖ*V_\S9WvvC𯝳viD<աyY+yKw(+ \J<Od9};EbK5J;r7voBqcRZ`dʁs&4 i%_$wTm2>^r)nmjE)NUV35c.QU<券J2)'d ;hj_O%`۬hIO=n:q7)V7{.nD`UK"D@Q |?H}"Ŭ:B;2@%8H@GdYOzįYnڒdƙ pK-jNʉBMRh&an9 uh'- s=F eQF9=}Usi6)M'M{S,n5_ͱCG/,W+DB$!է:6yTHފMk^^zE}csMƳш<H8aWz$,~s{QfV͔!j06ph XE1UU*OЌ ԃW@BgCA}:4,jp1%|tYrƤK*ihNRu&Z4q_ R(>vig9r1;e / ͛`^,<ǂ>kr*kAȓ R)uz-bUu"*Lr*ȌeUf"Yb&Rqqږ90vc] 卸nqv3|nx}[Z YQfSGa(lJxRa@76+PiKN`6s/v2= eLYKIVg>1M&'Ǵ\%.kyFgd,2AHBOK>giaR|&ILxC(軆>46!Pc2 6DGKܹ\LQcHi#)B+6s|L'!K&[;)xD#W^xFCϝ;6[x>4w;hi ^Rd-K{H8==o+•PV+ˉZpEjSZϸ:@\Yz+kQ@5 Y-q*Ws͸:0\9^#*•!zy`7u\J/f\ [kuւ+NWR⌫W+e/ӣ8_i\ Ar5t%FWTVcE\qyLEbV+[XĩU*3q%Qqoz uH;GNEܤi)Iiu5si&O' R$z[7ZyerN]Ġ[z4NZ4py{}uִ;[Oȧiߙ/?z:}8/ku۩hF,RqYܡI"wcEgTiܷ장_5Er}yCBIPAJjS %摛C Uh<ڝ?ڝSjWh+<4%o@e)D=ǖnNxzͫyUXm-T6b(KhңKXe *v]b~3V?56U#\Q$2Tu`^]BOX}3aרy~M$b>ɭrɩe厖L8;:dO?x)Dmw,(Z\6}7t5 K`mf^Z#c1V~0CM' l sSi`]yզxw@*7+.x u"}fq}y4Snq%_.i@Xo= U#( h*Zw_܄/UpA`UP(QVR@iA֑OYӦhI];zN°3e6#,WbFPEԎ #݌^7~˺%)⭶Vk,n~8֛!m#EotS,1 :e# eVe^ܙ82E=sB۳Xs,M-"::{g^óG /4 r,'E[c}*e&:=} _4 (^V[:ؓ?7l}Y>L0Neiw$ug}ƨo7'cXSKSƮ`Iz(Wۊ2S\;@Fꋽ8C+ud  8}ι{3Q0ވ`+l#T3z76R餞p W,3O䢪WL~&t܀Cĕ+t=U+Ղ+Vk']N#xIc;a˕\Z-+VY8ppЪ N:X.Vd/1|\Jf\\AOñP=l4PjZ9*dN$pp3v5=(PT+};LSTjJ"|y[_ XStf\ iMEՊPOtrWL1~挫$79{`ILSzz ֪i1,i 7YW+5y5,WU]ڗ+V̌,^UtEwrU5}WVO1*J4:Aq5H0}] k`~ZTM:Vpf\jz)ԗ8i<\V{? }}o=LSg\ 0Fpł VZpj+Vr A)bJ!kU0u\J 3Wज़ T+P?Aq[@}jtkPUj/Xf>;R LӬĴ]MQ% 6{r\ 4Y|#UNW_WVԃ++}-bz!#XHzXQkA*quҡ1Hg˵ 1Z'+V쌫W+iz<J 9L7wW hN= \\ጫ]M/Uʨ"փ+|-"VUqun| 썬W,יZp%Q !VquR$T5bJՂ+Vk&+V9W_WRuXR'h r&`d2[KAgd⇗ JWw[c{^GgyzA9iBg!?Ghe2ekbgz I<ºyݧŻ}xTRn6^R,޿*͛xVWѢle?P2 ɽg̯keK9vqy?{ΚYS>a;;a qzN 2򂢧Qs&,s2]-?XloI%HtKA)]S8,#( De $ +,y~]K1?.Nor?.}e-)dE.?`d 1cqM>h%u`4QCZ!h/Bi#;.Ge %( !$DȂ6!,SVQtal&]B-AKE \,6ʤіFdr҃֒X6:Ј6ƒ9 eܪ!1RB) R2YIŜ HJFdM ^Qj)k^PGN.(KMuQ*c-,]H2h)Ţ0!Lk)rK_zJB6 3y3ŘNSިdH&L8ߤ06[ß=C Q$%V>J! fvCЍU0Bțd&2H!Sa>I~B5TT imol@k&y)+ s@l!\%B ;B<\PnB*Npօ2iE)mK<%} ڸ'So$XІbͭKU+JlR(sPUdhw1 xA HgHx*שk ލA7w0u .Pj*~I~J d&M1 QK" RdK$ # vQ)]L.R{ l)^d*P֋IlDyR4ahBl 3hMkC*VPB0j΂FR"v5 .9BC')2ҷDW:#rLh#o2M@n(PlA<:9i裆b;i^(;8D1`,YSGM4^2EY)ṕ\V#7)nK YGG\4Qo :[M9&*.`jXEGзR4 snbS$DaWP<`"dB5ިJB15SAַacUG$ȓgR%4DHw.52R5Fi xnKe8bF-`MTvwm;nZWc&" 83H2'8_rHVܰ,yZj_fn[喭bm.IErsY\U cM5>g.J4h̚` UT l@fqB8;*Vj M e*B@Q"Ł.0͑H( /uW* qjz^i\T{jF J2H/k..ϡGӘy} }^U 5Ik ^h8(:&Y,"1۽@VQ=r˨ZQ0Z}pGФBufpOA5^H+fHN*&V/2I;L'ȃ*D̿hcrLzW-kv% m-Uv%cc{fG1^*76B7# J=$]N !iZ.TUFL1L: !'`Ge %tpΠDN+We2^+S+2]xqy{*O׫d:@Lě!qdm+} ,:T% 9UsP5"V1ﬨFy¶,9e5@D~*BJvzW Xޛwy2W}aH&Xz>ZbM=A!%DB>ouԆHTPw}؇`8B 쀺SPA~,C)y,a{}Jh vscAX A%DEHm֐h U3ڄ`1bP{Co,j PТ tdB\"(QIb$#EPڻ$?<Bjw7qVp(@MtkA: ܽF}E D+ (|<+E8P.(-lZk!W&EׅiM F:ro*QkO^(T /]\0Fn5 q?XoC1JWY/H1!ٱ4&͗&T$=8DԅK;)JN-, 弭T'3tuE#BPޙ#'TQfm~4w{qb GuF݅jgzzA t-O|K`N;t]ޭ/?>&Oio>N1y4H"+hfvN~]gɦI?,u\؎N~ћNs{r3=[^.J/ih~Hm+\\bIgZ͖{U.7NX[77жۺJKqj-46SS%FzW7%-9Cv«+16CWkoc+Bb·ѵtE' apm3%B,*Hi ]\L1AȱT *hCK ot,"޴BW@ =v"/1u}л _j?Q.OLW'{J72rЕcҡ*Xhn;tEh;]ΐVֈ.QWX;DW Jk+ul7oz,ܯ~oUbtE(f:C2R띋ck;:'ROqBz[TrnXd_YϽ ^ds uOd1[pw3}˼JeϻWWÃ{[p Q\mS0෇Mw ^khjXLe;NY;wU>1?7Jr1@/zZH}con|~%~;Կ QEK}ɣuO:יǮ7 J!Ť{M:c}FoBCtAS \+Z+BF_ eb,* =A3 *NWrlAW3>\'u`oO,9 D0c'o8՗|CtElah;]eR;7+%ת!"Q5CWmc+Biҕ^ږA6CWWVОiΑ4F&RT UYw{!h{ M̚Nӄiy5 ݟќ3<>y4F}MЯjNn~NI׵X<]~/-vX[\`:tܞb:'zZ\N>_~zK:j=۫zu9ݽ2"NwMaZChٍ}[WWHhấd-o?JIgs}ENO=^PʣK1{r )DEΘZDJ9Pkh}ݪ>HĽ1ݱ/:e}H` ˗k;+_]v ЫUw6FS72+|5A jo0:4kUYB+0}[D背HT9}QJ􃄢ٻ6rdW oŋ{08ه`"'䑜d2[w[7-Kv@XMUUc1ZnE'07pˉXn/u̹˂ x>]@4nQ;!&󷶹tpmo}rG7M6_蓇xz W!];XȝMF3UXkRocU[7ܺ}ں=pE]U1ݣv-n̴nw w>8畖a<o~u7AZ>!: iү=ܲS •bjtwK'nMfTPzYs(4_TNV403ninϭdP爐\jjͱu;s }5[xfIwxr~v=#o١2Usq`JԮy W|)ݤ낤p4Jc7ѕ #x`'!R9IFҟ4bG"O{nR^ࡠ^ɒْ3ٜ dN2i2ֆH(ۢ7IHP ܵOGI-"Qi1`6@$&J`tL;g?IgDХxAYgw\t g+HavSn`M ǫ s:7jl6QB("k 6H7@QIZ9C'wTH@'ɲ_tΌNm3OBV@A '>oN}D "9fHQr :PFt2Ť+pNF"88чd0z{Z`RkȦcٙ8n-fӳyAAU#?=M9|x-vf]ѧ[^Z /:.?:es:eG&dmހ"]9GgT$f}%3Q(SzC֐S)гqaTi)+N@H,)+M,|=Xygꮘ~A8" ~e՞Ws?Ӛ1]|L%~]'E60Oi,˴ [x68D3, 9HuYB:}_WR9?K!1b禁;]3afriFŷj]ͿVYoYJ_f9p>N}ghᗭ}L4)/ mJwsV#,B[To5:W<ˏ6mXۢYVxyP<˧qLT lJbv.uNѷ uHNzok;u?ϯuv3= Km]f~kP{*v.mGdV$+hx7KL" ˘-zf2d)+gI~BmLrw5৵'po&K; j5 #4f%Ov Z·ӻ_SpWߠOy li%re12: rTuEZ|'^C6٣>죲nQ7R[~ gue]\iiY!)cQ٠ G=z^EYH*}$`{_{NBkH}we(VM,5.#L=/M}Q3-AR)mR"D_|%k03Jm6IRx#Ri 1)u8#m15$%D(DN:L 1L-1v<r_ ۇh;mݔ$6+^݆4= #QɦVq@s$ 0X+~+RF&K`M6? yR 6(#i=HP$/sJA:Q$bܙ8-x#z!kY-Aګ=hSLKm=<^xjw3?3{o&4f]tiRنgh"HL-=<TX1lTe2TM=n۱Jvq9]8oc1Tu y &8Y(xX;ncsFM_o[YT0O1FDJHF̾|AzK9K"Jk<Af&`ԉ\Y1UyM"'ϱAD$}&SG^?̴gKL)Y:HR>ٮxicL4X%jR"F+K+JDYBR(B,ٹƔF_i|p` hk~3w80~BGn@V<'8!1F8&|{z7`g (\ L*EjQ!ڲNВlS4h]Х`=zr Rf k;b6/[D%Pr*%opWuTpkլ ]Y4Q.I+41T WM[vRo,;Ss+`FtKPr3c49'ܘWl!& f0.9׍aso0jDhܙ7zF^l)_L^L'>;CP8ML49b11V90 "YfQRV'$9Y茎u'yg F{;Mx^YgUfۼDG++hK#^MF*%uևshR2>3ZZ15O?80$6^XAbcPGrڧ-BI{A-ڤP;*L XkK6thr62bƫ Cv*3b(J0V%Sל3qs6ǣ7yu3I M}1O}f19/s=a㛗]Oxdz6T9$ š-Aj\e.Jnudc3vb3U7{vf_mg ^ bRFl;DFt ]+4owL*Z20JvuX uX|1N՝勍ؠ{W\8ub]: |w9!jCBVP:mKX^F\U]u*3x7 ThE)5+C{_^-cgt :~Wn0cUɧG75,anZ~aWNf//0y+&n@ u"dw=Z&;%_?.nWc-@i gl:NTS&HA@B`mN'75qo/_ Eׅ{& ?h0O> ~NoQ}*O! Kkbt\r/ʍ\{wIh:!OpDqvy9 ūWşR*KU(tą҄ Q'B /gum/|-GdˡozPk_ "Xj~rntlQU_bK׈]n$oGCmC5jL_S Şwu3Y.3cXeIsK,{$Xaw(Zgu8m<7y;]Wl_W'>$qY q[r{.焇5\/ח@{>Ckw%םvb6PDή;Kk+j^\/·DXŅqIJeTV-*mO2NM4[q/}TQ;EUZ GnG/VϨDƐT9'" *') v5ЧrӮdmn您3̺Elߍ"ܲoRxR@!J\W\F ̦5/糹6&ZCR) 7p2F\Z[%\Uj*(KMC 4ъ8#  $JV7&B0i1LSjc 1THYuK:Qur[X&灋 Dvnc֞V8AG:.nxs^U3&W+Q)4A'Ha`& xmj-9ÆٝyjGGjGc+0SQlF$ KP T |9*!v'}4 2hϣքHbpF *Nuεک_GY ;cǶU?v%"k'bO;SWdA+@%7j +/IJlD8JM(hr 7"x4;Mgo{[<`o\t*|?ޥjɮ\-s\xg:E<.>(h)%YHeE$B{o37&$\ ۪PCa,Z.{G8qFQ;> ~)߽pj)]IW(xcB& Ԛ U2qťPjo$XK Pе%\Zk+PiquT v[mLgnr7gWUgLh= zSD$^?_^(f݊OssZY]kRrg"#L``I6FJiTkE1 *=R1esbZ3fr; WV%4=WJ)CxFBZd+kY6}`\J{\!W2Q@fc]Zy\J[WLj+T' KO52\Zuf*E?#zW(Xc]\s U~FJ$P#oLUh(ᱍ2B 6TٱXI=v)zj4:#\Qt_Y6BZ+P+ :Pez\= %pW W? W U޺:F\qL8uW(fc]Z+dq*quT*a:/KJK&Q"O-# h&٘ͅ \ Gs7䏫\vލq8 LV?A|*&x+S27u&ub6dNkН'ȿ!.p`f6Pu3shf>ڄBxQ.fQf"ǽ/@|2{Zmފ(H^H[ uƾ}czo ;8<ث9/p:w|k2L) ;oՉUJ*[V"$3ZM*NaUŹCq JY(p2RJ NL=Uj7KyR~`xURBDJyU gUi]Lu v W霾ddWlWJI(0'e= k" -Vd3 %2jԪZ8¡(`BMFBVe+KWV U~qR&(>䲡|@A:B\Ad<'\bPw)7]CAԪYWTv Wh]R-W(C*s=P~*qu75M6BV+x{x\J.{\!"d+L W(\pj:PquTnؒ._+H O P-jpߜZUr Qf՚cTRzL!%+ lJPK:Pe?fwRh E\\rP=WZp.slprWV UqecugFj @$Bb8F\YI9b|+{LpJIz\=\=;%R(sCoP>P{#f?*k+jע0I3>frWVuWRWG+59)FW(W\pj)1]T=W\)s~ s-\\uj :@ JP7FA* NPTch.noffF77w||} ooN jԢ{?5u3ǽ +5*Nkr&} ڿnuR]':.7T歪f1N"hʌQ"SFJ7,Z^uSI_A׽㻓 Iݝ;>FAX\d4}.>gnd?4+xYNRWczzn_: .gMP -⼴AhVy='zh2BH2Q-T+:?*GpTKjW X|Q. Ju\J{\!4ge>Uل8ZTi9ctSBیp * @Q@}1TNcW Xs P0 Ώ]JKz\!pu> Qչ F+T+Āe}FǮUU3@jҲ% p{\ZRi{e+{ȍԊCEnkU'f+f%1 RBseB:B\ V\Mߡ%jZ%V|jFmtNVl0r`R*iT]koIv+ ;PǭEMand!qs$z%;=6)Zc%QأK"SU=Oם{:M6JjogDWL3Zc̅ut(t*~N ]10"h:5;FiTHWNoqstt^g͎5b2hҕ@[]1z?׮塙:]} d]2+k\/R~ECWȡcR$K~p3/Vgvl+tԡ$]1oTtp ]:]1J%ҕ"m5_KWl}_>{uB:]1J%ҕ&]1`fCW ]1Zѝ^ ]w@n!͉7 Z?z7x=@ǯw[V3{iPٯwn<<}iZeߪ};(dr5Z1{ 7#'1 7 lkо~iQzח^҂1]|{.̆ns+u*:]1J ^ ]Yc]ppeWW@tWW/:3Ax9gj~h(<ˡ+|%b"ꂃ@I X$?|^]v>5q eݮ7G_xq; W5^=f}ɉ9o]1_'+_{H>cX%\]]/P氓*og 9zuOm'N|M$}A+$ g4}ilhBt8c>:pNj߭x^_V^9J>NwZh/[# Z~^ʖ ~S8B7qg[]v;퓹6Tδ/m㮟})Y%V`%plk5FٵڋjVNf%jut(t+ȡ7Agz3NZkI]ԕكLR^+EZ>];]zj`|;Yl~ӳaIf{viqZv{_޼y dos9hګ]@m(Oo{¯|12 ?*uLgc#Bo[42-֫K@g52;|h #c?n |voZߜ_џ|);ŗ&>ޏvF-fՎ,{H Ƿ񅬺ĝx*۽<> Gwϳedf5]̔ -'  ˑPOۿ?]j%ī,U~|Ob ?\ؓjTQke|z]E%iS'*2[Ӭf5-ܬރPUld7Ǘkmc\^-\V(h%)DUq^[J6$MPd1 דߗJ1UB"ܦ2.TT.Z5D1?O0vҏu4JF/wBC:jR& B9PC8B[Ul"IZ2 FRR(OvJX+"TyᓗւQҩd D)UЌ\[A`uIQZCдo *TcԚ1$kFH)Ldit2Q0URB0'ZK}D,=XɌalJYhl&ۀ)J#I>bg@DK3R4YWQI !Sm ӜvF4**RPafcgOb15ZErm& 6 .XD!}z]F D>wg1ן7Yxc*lY(O`ɐ>% V+s\ֆ(ɳ*K'jjdU9挑9QJr-^4kAr-;u}dSa4n );Qƪ)L>b##+$V 3_G H" 71rL VؔjH) AJM!!S`6ՖLpRTD@ (X'i4/f$^f!)ф+Ŧ] Y 53,HXF Ht=+l )D!; D{%xs"$vTJD|.ːVAl 0`%'`B|ZCۆu\qx(& %JR-*jl_A2hF'-s#ژ[M`s8HIـ0!-CHL%eOV>0q`qHJƬE6,$C*ZeD@Ww@AQtU:Uc`q^bnJ Vor3E ] * $(l`Q&T$2zI2L d>^T[ЗNܩmɘ-u9/m`-$ >:FjAuPqLJ6# JF%I-G B˅9Y<"Y78_4)X qa-<%EHiD(2 !YiIQta,9Ѵ<=ŋ`C%!YZ{[x3$nCf6s*Yȩ ՏDE}ygD%ۊjLj&@D8X O7~ruvODly\E4Χ,}'XR Xk#tdWHcѣ.)6H /&sP6D. ˅"8Bik ) 1%%@ᡔ4Kq-Ah' y.!z b@ l+ jw ,hLMD0j!hk jӣ`EȒ,B\v`y^Τ$1S"I/&Cd@@GƢ""kTA WE2$cB8*gDH!(Ɔr")5Z'hV =ipI+ j@e# o JP)BMVD鹿*6 2ߤn$,B S5/VmkY 0dJ-S@h8Xq>zy^:_.O`x.^.s-W&wA.nn`3 58)(}˿{B bjSѭ1+}0DL0IJ)#;&@j}q*qPb驓:HQΛU~DZ+򎊷V}TQn{E[}7Vͭh1uiKE/}TuI ^PBF>rv|K`N%x-)F~pKvÛ_`)?\l$*Kh?xvwt&>Xz>]'x7'Y]qWrIh~{puW|ӽ%''w_n#жoz>C[/֋/ɟ =ٕDDq1d87~D9B ^t' t]<֝@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ ԝ@ uJ9pPq.hr.N Fk;s @(NNNNNNNNNNNNNNNNNNNNNNNzN qk!9r>{1\='ZsN wu'Kt0NNNNNNNNNNNNgJ{H_!e}0{z0@v6<-)Rj{}#dQ%(VCUŪ+EdLLLLLLLLLLLzL 0Ϋ&P ;#&JGygiRLzL @s432(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(32(321Z/ɇojZ]ׇfp}PۅU} $0_gD\0JRv.ĥ$8yRAĥtgiVJku6 UXJ2v*I٨ QW)Ѐ3RWI`}>!G]%i9uu$種j`x.@ݫ:zCoapoH.f]_P̅Tůn7`#}6j:ϦoSWIJ&~jZ`9%9glPeVueUR#種OH]%5?urt6 4種4I]UW 0%磮 J ~*I)uVWQ]i)~JNRٟ{*΍Ӵ)ˋ/ft /o~܈ a?V*`?DZQP0|b2&*Xv@K|vIhקcsq MoՅo~~v#͆V_E O$N\DaX{d`T0pGf4.$,RjC )|;ٛJ ? =GT9e2;;I@;a_S?]īYAotL \~t=f1(?v_R?_mŎR+'m*Y.yHoZYܱOxs5wb JɅ )+&~)u,=?^r-" `BKB1)rw襵lǝ3l=\ 3 &.v/?n߷[ѐ݇bύeKotw/"7o_<9 Hx7K~qmEV,;۽R4ir4]ƞӾ/! 2t \9֡v_g1Y[D薘6V*m~>5IݖºƼ'k _w.8I!_<\d)X\0Zp?fZ x F_~qS_¯uDm|ܰqwn95vRkmf֕||dH1c !fqYJ m:SDi8"P(caBB|p L}j8s;Kq&s(Ĵ Ep}QaGH~֧RYP 7Ce@3*-\Nʩhj"b* f-RтC5 rtʐnO <H,틖ƌsf?o{!R9@HUWbYI6 I"iLy:xP(-َ2ɀ͹^˕8L\[JJkۅ՟(Aק[_=^3,Ko4StPtӚ<0bxd,13Qxo$#&"<@날ܱ+ʫx\qb4_kӯ6.nkX%(YM}C;{[karg g6 |6e7'dtrVLLuEqVu$F C\R}3tw` -׽ -[CU*xO#B6IS"Jj:A蕎J0m{n֞I;-]f{KP-_Tn☣T`Ҕ_Nn>[E`cJڦ1g˜qWa:,!үi(}e:/ZjoSP m[[|]uy=c Q_5dd]Τ`g[nUn7U$ tS:pv翝Fml%ve]%#?ծU󪜮:cv6?ߓ7zR',z}Aױ֐v07m[r 7|߅Ɣ"s) {خ!:B`@Vi11H8"A;$ *}~O`ޫ=0`R /a4z & !z]Bqvp"CqLo9}ҦhmN,9?Ўe8{ 뛒͙=ÈkÈ4:U98Ds`g_޽Ppe}]O62P*XhT V0IILf:+YnYaQd(GaKՖ:2CoT4Ԁt6=U^VqY¾(]mGl`+EswZ:>uX^ʷ̙OfcFGGBGiI~ħ0)vM JiI`$ 6;˜9r"u)oWH3nHyw`YfqFrx?Tffظ0}0*XH%aQLn k,)PZe:CG@&kΧRKrje 3 &kO,ZXJL!0"+h3'p=sEa_ V'J Q|0r!XNd# d =!"}2E4^JJd 1Y,-,xGI4 6jM th|YÍ"qv#Aʹ__?5x[@f{hpYI^;"j=*) 0; J<2Fτme֮ԼnMy2}z7bSY5K(vɳfX@ZQO!He !z+[k n$^"e* H,ڮịE 62( F刊`) NP*:dZ^}Y߁j4N⺴{gr- 56E`WdGؑa<ڷ,}=?ٟq<ݪ,II (z²( Q 8[$ 98ij:w a)'*pr<*V tjd7si3x97لxvpծ ݇ yJ`tB{ L[C,.=$0& "XpH3ȶl8.T:q)%VbUK,m@h !Ѱ=jBruKcrTQ8UXRF":xLR5vJkl\⡜(3h1W 5O݁Gr*.>;p\Fn D Z(H$pC݀wP7 =rp5$$(A2ĠtYV"Z!”StZ( $rhϵ1צ帗`#mi#9% RC`s\&8Fbޕ$B]lh}a3d-XcdadYͮuUuV)4FƐ:>hªq(Wq'J@Z# 3MbH툎@!c$уm GiY@H 31H [ $ɮޘ-],=1d $h 4(s. 1(HqIfDJ 'Fm'"%Tʢ9*$"ʮ,ƺOtjƿ&T-d' LPkMQ`5ݢ޻t)zC.4˃ :8ubP ;#9fLaoc1")ʈA%DhhB/6} l} v_W[}6z?mZ׿BC|Ac #\`$F a#đcxd a'xf:`xKȃX`2(S>ꀌ(y %Sג8ƱfVڠoh;K p7 ? ɻKuPhƚ"784#O! *6!1UXhbs `Wv^*Dwɹ!<#%1A0XIg m)±2!Ƅ);i`lpoA qq ¢Sg&i3J0) eQXC*X"Q=Enշ0vd}XKGCu/mͮo -KmxZM],E|0A6Dxd^&19 Y#, cHTT`JTѣNmƷ im }D'O>?z؋*3<𷟎0𫚤Urj s"D@dM9#vLYLn;S!ir)'h͠,_,U[UvN[ }f: i|FYnCUT`+U|CUl[U_ռק\] =\tILSJM q2USg7_M/Qqz0of`R{sҌ>/}3_b|vro-(>òݮỸL'E s-YLf]޻wޞ6bTU:df?%iBy̳8pB mA a{pjʋ0¤aFpvVާ,1ae׮d?>*fH_E;ڿQ7fC0/Xᩋt:g%i.y*d| Г㊛~ҫ2y-D%޿^,[lΏ 6݄.Ţ y~y@5"/b(Ii vMt'yt6qrR?)_o@1ݝVa7[8bqWDШn+(jEﵽCamv7Лj$,K^Pj={CΠl6=&qDUI+4SV]ܛw5,mX?:ζ֝gVLȹH r3+uYeJiV"qlז:켊1{UD?Zpx *3JS"Ga('"̢$r%𵀵ā.eLG5*ZaEb½Ǫ=Fj+{RMDTPhNI+75J ga TP9*g\qO \J-]X ުnK=*E"#(JqtOG& rQKEUW`B<UX-b3jX$%1hMoŭ5r6 *\ d K{R硙‚KF5IOqU 4F(i 3]!ADv'Pyp,,S{+S9jk_lo biS%R}-5>VK-kPN_GE4aӷxrRd5hwvҠk/޾(ϭ!UTs*G:"d:2\Zq:XHQTR@Gk6(&=)oW12#XL9(JRm#cklP5Y5,l3B:,<(,S6&xY&/R~-dP.u A|-0b c\rDS0GK@*RpE4pAJLj+ ۍ㢊F!){D%$ؤB:`^e)$Et!"'Uˈ9Cb9KKܱ'm4|FULZSQ׊E qP*h bZA ('@p R:HcZYR?H7rl[-"w!Cb$QaeD` -R?,CQZǹrsBڭJFHHqbS#/ gN h҄Ij$fXv!٠] :2p5.i˸H:\pq%y**L{;-Yb91Hg eŽöc[^T}ey[jLaYozs暜(U,eжڋqܘ}-C}r?fM3RJ'kUYٮU0T$ΪNT+%='&Lj+OE.g4h-nCeq,_mQYZ 0b_L52J: ? "u0~*\O%QȾ$*9T Ê31]<\|̮  +p0h&,?d7tJj=4DD33iY( ?h:9tV0#)6gF\6;˜|8L @0&`L \Ŵ|M J!;34-p'W@D`*ɡU}D%\=C\ltpP*QоURw7$D> ȥ`ΙִWJs+AY9-`A \%jw*lWLts*rFLҽ6dS/bxAYvRh-fEQp )ΒovyrHtѧt' 'z>zȸmW56Ou^(?Ͻf/BB6ʏ^Yb`MQpżBiu^SNHy5(EQo)njAEr!@AѦ磸xr.x )RuA=Q8g\k]{oɑ*+A2R{[`# V)qM #W=>, IM)25uUuu*58gbSSNA:I g(>$UlX,Ű3)m s+k3M1[KTEFcD2Y+냉KM-a$X0h!eLD:sl왫bf+Az^N: myA8w^G1Ҋ_/6T*._U\0GV*& @h7>p8r AgD8 "$^n g`x` `@bmDʠAH10C"|L7,Rg2DOLIY,(SQ K4flw.I)}]&E2M^M-vgt(%hƯ{g*7V9DId'PjG M rNi`Uv^*D'Lf;V>z<CΒHT|tF`9\"+R !iL%3dA7J|^g֧ό[AXTy$kF(V~P& Hfp/`Og,٤~X? }n=2CqH.v7]̤/$m]6gN $JiC}@Ge#5²88vDETJ=:AE|sӁ9ѩ㱌a9k֢:Č_=d<*>j2wj +f~u7GD@E'~+BLJ9잟U(:d5Lpߌk)K FtG~O=,ST^0)p3Fm_ 175#3)ܛUxLWgw3lذV]Lfc?Y7 6 pF .b>`#4׿i1麼yr7?M{L^\#k*;l gغNX 7/xЃYdƓq5y)hN~|&|p5wS\FDMQ3mz8   b @6\~f`A!ak,n<uCs> .b/86(IK/4Ry#`'!$9_-f57=sPMzn_ &OET+[-uF{ƽ*'1X uubC8ɇѐ7}'qP3Sݥ׿xBϿxכ [_.@LSz-S*xwdZt=wGJ֙+0M;Ά[4x&oFDWؠg$Geeg_UMw{Vx2uŬ99gΡޝv井 ϪDym)9wiTnf.2Ti7J5Rǐ稢<ĢUik-Cx<ǯy/ZRY5@4'K ^Ȍ ^`:`(HFlF|J6,ba{򌅟 .)J6_$͸ܬwnwof4 g#<`K`Jh H(3K+#/{* C2'ZK(IItY$RILBDN݈̈m'cfb>K&8>4 V1iAMAcG^+1㠡xΘC"l9,Tˊd>*XȠbVA ('@q Z:Hc2a6rvamwQ OF""ɌgDz`TXj&Q!伤1Xep˩Kp|q\ 6QcJp%#1)DXҠK0I̥*19>td,bŏ343.3.qq+̹SX1r@VtNnRDEyca.8YfqHsǥT;먡镑#LjLSK0P Az.9 mB?z-']e~}}@F 6:$0Q@:23#9˄`njoXilYju0AQ񏂜mlO?Tn9ڞզ@"M Lz5.W.RqXtֿ XK!e,!]Znu O{9s=G C_wQzr1^:N'J.SC4A0ɵ r uA!gh(hA+M1Tk%iSo"ͥcy.2§(oC1Uo¸Cl>Em/}SOWɛqFjԸRKѵ KAZRPQjvY\@]w!pU&&U%ڕSQ#ZKrz?yAQ{(3vRod\_L@j'!R `Ɓ&kq PiehhIh)ǗƓ7/V}]_\!D)ʫLiC"+/`Y\kоKr]t[fv:[isr?|3/GfoiƵg9GFGqH e䒲f+UtE7(_7a͓c޾@o%UŲގ[djMZ+*/^6He`%Ww|{~hd⪷bXU0{73ۈÒQAh= ?$7oM"qOW\%'ui(00(Gɸ K}2fmE(vf\Q١0tuɛq&<I#zXvIS+֭]fyv3WM;r~?\TbҔnV»S`a$[NB6dθPKT}ESi0ry6+j3}dj嘗+x`h{ ݛ3)a} :d94M ~B-)m+wgZ{ؑ_0!YU|`~p'Dre;뷨ee6#nubjTf~jukM6R_|*e?XL[ Exr<3?<_\>9}1].ϹNo%mma(ӼXҧ;.R:}&Ⱦ :*5i ɶRiuU˘P$-t bǜ_ϫ*LY{"[ԤektqBSH**AR !%(MݼhU|Q;ߣv >uSWmr\=tm jXKtD2䄩:=EebՔ2nt= Hr|JgNXDR[U|; ; 6n%n{rсБR|t4Jlp,JZ+7+_R(p2g [`Oʹb0cJQ`)]鰆ڔiDE)]uUG/aAkoo|r{{{ݲd2D{8@ypZ@d bBq><"{%QT]@YxK(s:P ekKeuАV NXuZ׍t:92c* a `L&yuH[#:iM~u2fa(,j&Dm0 PklIšlكp蔪uƃ6X*PU}`bnj1gPX\;? PDXyeRQy\:F5VBZgܽ#_RUg)Ϻarl{A0m-gt+뒣**fE'6o(3t&`0dàv3;3+~qGR=RhmV{m/+Tnu1eu;vz\5*JT^RGb]tI?{0ElqT0,!A3<Ay3P #5&YjR(T Jި̟Vu{KVbw&Y}d&9"xJN$qFАΌ Kn̓qgg܍;gqOaY|-.g)Km( aNXWVՁ="07QL"1+:I #?#Eɠ :~0Nz:Yf]-VuŔ3Iz mu!GҤ"tLՐ`Pi=:W%K6%JY=u rV>z\ cLtPNPJ5KU'U+Zh4R=E~/|=n@[AC:@@C޻[]ow}QdJQBZ&s: i=ZS~H &.΀`K-/JXCbw Fq[teCҐ.Js,&kVL5F'.tK24(*Jކ L&B R+['1Uj8%J6L3s:@D?KT}4 Q3s蠑6blxJ!Y!ac<Ǡt+F{`PWcPj:"YE=jk-_*/u^\,8?϶$0,ޙM$.9CݖsJYp[/{ G#]Vk:PmQJwfA&ʄjDy[&O-N K7B;hv=Aj{^ =t[-~ۗ׌ $r*[W B.%"m Sl R)B$#o,j:yM||z׎g?>sوj;k ZjY>je d7ǓOȜOSW@w&뚳OWMP]9)LHf492o'w4v_Zxf>׷C0x.Vs^ӽ@Fyc_"~G#V#2{xt6s= \@:)0*.W_l@Ku`8 S]RFL1Ex`r9QʺHE>()~T33P3:-ͮ\dY![MUڳ[]\&Icgs<$~Buy~y>2f/{w)[؛)pqqy:|Nӄ1@_uB)LS(-j 6[%FtB嫎dwI^QT)^)}PWC p:]3 6k=y==\5+{yRU3ф+6W+hݝQzJwCwv2P1mzjAGA L{&h{R/+=Yt>[R)F&[BӁ XU ɗ۫%|Y: @|rJXpt>+J?Ҿʀ(J?ʖݢt G\h$ecUHF Ad =3N%r;+hߍDngS\lM{+ۚs%$xJ'w >u~?fٕ۷uN薫ũ.x>Ԑ-p+6ocyyi&ZnYb!Ah /5Z 9)6㿖?M_t.Q2kh}_ ӚyLE.5\LG 8cT_~I{A9LҜ.:o:2/PbXGue"GpigQy[#\~x˷w٘xV&UO|d6C¾k*3j=w)>|ͦ_޽^ju_}hf 9yH^L:~| ³yuKlY֖HZnϯ-:r,t麋Y+P{Wӫ.bj}ۇbV$KoҽTE?{m s24nC9'UNJǥEҤ~CR"%"U1=3@57DeJ3qmyG^yd4ki I~Lm ./N>kKw0Iڹ߻ 7<|1bV"0>y[.Fίv'')oz~}&3~_ߢ!WyaU2vTm,݆}gue$:G=v6{'HV@^ Exr㼵 SMH7AB`FrѕQ>OzbϾUs{mdJ);+$e fjpIHf2h֬Nb-A@K{)P3#l5:! qR5q^%(;m7 jZ]ī\V0tgحVcj;tt tdbfo~SCFAFv_U\Hxm ]),[@rz l\~ߝCdzqػ:M+,=[6=d\Jvl{wl4m߁7UՕony8֭{ܺ_+{m5uvH0HKq&AimlMzZEYnUVe㨙+Q[Liyb4@XדߞHkıg ,RZe 3*zO-#b|~a6:;YUwr*4&0(;]9tnZ bZvRekRq(YGb$.ILt!%6[t7!*@BŜݧ8F*BJt>!J2p!VQwژE YTĨ&5 FB1JŎ(rsʋ9 ?}m뵊ʙiX9]11b:l5n3]L?f{fO5r)&RlrJƛM\`Э# gr~8||,q&DDA[3LƤ1J2(hMP˨: F MJf *oO< WEә!z<( Yk#d3߹^>"tɟ:AL+skaqϊ](_vPTUn`nPtx@xU)RCj<͢[WJbh<4hMV%Lzޥ~Fx Ay[g$2*R\rX2e9ӐHg3 \يVd;#=pFG p#)Y8@… o lYMq2d<ǙX-qN1g8w\*"Bʎ &4"y^usJ3U'e)R0%-7XPwkREx%4ͱccǟHk_6MA;@@g ܀ĭHGTLƪ$f@ kE܀RGD0)3X ;4F e@I>M F)cmTMA*}lDSqJXƟF-VV$cLjOt1('2etv2@6UZcc^[ѓ5S9QN*.1Fe K`~`T2].&صK1-:F<QFQ$&y邑RsP A mAd2UAn&%'AxeN"ӑYOUy`4ؐIKDd:m7{)ᥩ֋Gl }IPsX ҧh ˁ uYV0A l.΀,z*R*N'ǹvs$xzC!(lN#ZzT92_;L-qqp_M㋼-ܫ_]~[8E+:G؁a /tzetg7~6^ LȀ2 AOȦ%A;;&&)ľt_QU\'}}*nA}pO~5?ٱ,q1@[HjB<%51:Mt8Ub %f'+.V:qR鉂5|dXco$|8?zY؟픦kKmGaDa1~Qַ d}ir1@`b*p R6eYBlG`Х8h?7 ^%WyU`5Crɇdr7X!v0Θ( kNPc&},F @ZΏ &[۽89mkskax'vi/?&mv9IKVde ᥉gpiv³m[6Ul40;SG؁ 2qܚmLAʌYISAh:SE㰌qs70 ZQ9^Z@D|a3'$R}&!%heSu@GnWG&&};YסŖ9PQ80tPE / 5,BVK5xFwlW鐌:/112)@p%udqVj.16#hSZW{OkyZ4WۘK!&~4>x9-A@ou{7&7X.添;]tub'?i{ϖS-%I,mXb iѬ 1TV=$jE-aU^]EcrwtU3h5pJw`47NiiD@xQB@ȁP~߾Sf~K9?Ws?!2'=~fgVO??U譞_ѴODJbIz{Ř,t7|d,X"3ExwHk$ % @- +zW5ˑuV頝hjo %Lk-&DQ (CYh˾&qN&˂8;q(UDn=im*]TV7*^Ԃf|c RW"z'07Z  aTD:&OZեP؛ljn:*+r?>XgoP.ꞹ234gD 4 VQyl]X: G\ z!'@#Vg$SyDf$8L%$y\d֛iˋXXUuAٸR |, eA~:IA@/ 2G9C'BDHF5u*\i) R Ao(8FrC8 [\e5@Jջg\H5ݣFǪ@co@шP+ L% .ʇ z+EV 'Tn[sG>]z-wwUZ[u]yWLXv`r`"uf`D7ja!߼r.qQ?t38y~X}7Z?]c$|f6V1Dzb)}zmYC؛ѱ' )9l}_/nk.1W9b-Ľ6:2]5֡D2c0A˨ n;x! jM w-x~-n1[>T8;fNjV>]<.r>zʮ_O..v,jqJ#J:o͑m*NLzp?/~_{fzq~Vqoޛ)䮦{Ϟ@)5e>Ѓ W(Ku  ,k]*1-9К)5Xl :2{_Wuum,Q =BJU7AHC=!be{ʂ ߍ!&*M&bT$sF>]{W zvvU;\C+6r\Y.#+s~8}ݢ W8OGMRWu6 9MU鶝!B?(.m;jogq~15}#rI\`ۇ6EO@=5ޒ[xޗW@Mul IZdqQъrBIźJk~pc8cgc;L;_{Vd{k^rRPL.ք 7w甼oG|+Ai,٭y6"Y5{mǦ]~)v^ͯP\\mSE;XWĕ($+,IW,Wc*bOv]߃WNfkKӜmVo{pk_;xZH:ölf1ggۺV-PPy߼ҕQhJo _eý54ydCzmFyVyG2|5<9n*F]jS@kDT+ j=cI^ܮZ\7GMS*(QE(E(jT+(k1*>L(rQZ`΍=@*90: &+lT2b5vbf\MWơs)egdprYgffY|e\MWdW;X: X}Z Jd\MWkcqE @2bC;[W錫 JC{:bV'+L*"Zq*ϸ @ڪ2X=TZΎEG Zc;m'mY nMg} \_>6&R6Kt: uj[qI: 'n%LsGA&߳IeO(oxMXH&&Z@@e1!\!nKgOWv=Tnٌʀ$+cR j )!:Hɺ"ϱe/\\3踓q*s$qn2.!\y#+k IW/Wzϡ'\T`3t9nr{$z7c;Jg\:Ҹm},HD X5*ǎ+Vtq n,;v\ܡ{tSkq*f\MWYARǮ}Qbǎ+Ri =YpgQ43NO=q΢Uh 2!L`YO &F f2'i.%LW,&cUZ+`bckq,2z•NtVDYʧ.uT-3j:M)jx76cR+bX؊\f\= vZ+Ia$ZP$f\MW)Ű`:aSUwY%䴘+s{<|쪓`C[Wꁝnjq-Tm*vf\:vq%ګNW,תTpj;H:jR,d]`1\\z9v\Jo2&++BJR*H.`2"VWҫ)/ރ5NPtx|}H;R>f:91hܔtj~>p˫/I?͹GS{s3V3Yqo^}6tvCwcݾ5f2 1Owuoo Y<'p7;+ 7?  uq-\ގL՘D÷$=[));-DR|qc!j.EzVB%ٴTa )WfAmCB. \'X;kMβU6쇅kk0 a&J-K7/mP(䤷1gU"m=)ްJ\Zb$d1 (9qNP_E}^yj-pbه̼+4;\dr\\\'ݘrWҗ 4Xa'D34p' jhNHTcl|h MfIb?7meSPi0vp#pVVbdld ԎFV9mFM#dNw`}P # A04I@~/_۳\olb1Q=u*L%;KT}g тOBr>}=oBUe-GpsL mh dkKwR<$Dn`vs=!v2ΒcO昑((E?cEsƂjڄِ!_*X %%$ښ/uߺu)LU'k(bŚa|7!'ȃCK 27gcPG*X 8̒aQB,ANhGhե ӅY(P!N|0!J(-8bfRm>h-i`?ЀGtVp0PT"HcaLyaC [Fh#kYDIփT\ZA6t'Lixu bG|Vj;x{ 0B꼱*Xac֣6J1V/ȡm+"]hջޠ$0*!yd[Xb8$؁yoktuƆ;"RIX;#ĦfC6U6>B`AV Q껡L j&6n|#rgd,q|XE=5(!ȮĎ;PfH57]P?;14m޺ DcHnzPBdBE@i!ËjM:co-$*X,x40MS1JPkICFh:CA T&N< pY +TLgq@Qā.(ʨq` tfwB0 ,ĥi&3fCC͎1LDFe9;_l?x`g﮶r5t&{U0طGǣ.0mFL`-d3x:p: .A`"<+#')fF Bmv 2:Σ&u.E ] ڃJH&5 *2%+#`fK`%`|OG30}fHH9Br[!q*p },pP% 9աQk=!Vwt<&[WANC K68;70("d&K8r VlASе2"Z豨wPrIn} }(T&w}THGQxa) Xt* mM!gt]fcvl ԁ>:A:x X-&wP2 Êы灠08g2ysIvy`.`r6chl6M &>wPpljq(XuuLkv^k159τ5rFl bev}5\}AeF٦FqADy ؓZNFB?Mt&%\!ݹRATbk*g'@)a>֡;l՛Ͱl+4Ou'XO: A-$\ C?I*aY 0z(+ζHR8QIUFH ^A,k~x,)mL 1ǀຝ7V "su6@:XzPi@Hk&=P 1jGjcK> uhd)uDylƥ?qKo%_;%7qg|75M(/" bd;dWIG_0w_ l~˟^J h\9Ӈw?xs%#yate:h.@5 wT]*䙡꫺m+Pwo /<_8wŋune,on[uِQ"Uʌ2BoQ@},ėS JjBS+ WM~,\Q+\ qGDW%ǵEW.M,h:i(ui|teF04npdj7ǭ(c T.͹g+ʍz=]"Q͒cG VǓhM磫MOouEW6кEW;4cvځE2?kKí߃`zɥLY97evk?9卧4=)657Tsi&MK0Lj$hUqN mjMaʷ|'عqb֦Gɿ_OgKO\=FW~&G6 Z#hfg?'`!`.y?P: ɐ.I&X:9iQE$d/Iw#(#d3 OoN[ɍ]ɂ}81C؜@87z>zuꔄŁ[M`oMn,He,+޿LU3AR?l!0 !e#0(;d}~.|s5I/}RB]GW7]w 36gFT~*a8?]}?3FdϔquMs= *si?auFɔO7]`c4q2hgRfJlL+q)(&ZGR _YĤ1k &l0inR>Z ;Es/ Q| HJ#?8'OL&۔q҆Jr"`6pQvYξ>| r`}Q=;s8nϰVv(C`AcPJ/`P%"rlw9Ї`ְoMfOtyV.`Wǽߪ,pPn$>|HYӧ\Zv~ЮN;(y7ߚaھg2'ߏG~ٞcLr:;Y&mk&C/'{z]_=&Sϲ:;6i>lv9epPZ_;1' gkP>n9;w鄛qw|)<;NGyR{/|kl5Rx&LO'@h(8Vbn(egr0O ηȜKhfH"hv& Q3X۞ie/m,iGM,=2if~zZp =)%;?-$wNExg# *yWtSAk[%Đ|1lY}x]Ej$Ɂi$1 :Fb6Qq_d$ZȖ#<|BkcYhl{o; o_`p)?[Oo{:"JCc4O9Nlzz{5|k>U#mQrQy:m&Aiky8P1\ pI\!Fk\L?W)psXJVrzz͖*ChHXjt JƂ?egRٞل{!& kw־fV.xRQjRA6ʠ8X F n151%VtQ/}09E%cۚ58k6I VnA"w4'L)59//M[^K ɒ 8U(p[HzX,%i(wVw)6仴n㫧w! dRI[Q<sO|ۧcQ}}s9Kl/ZP8JP)h4Ľ 2z 9&I^l2dxѡOYZ%lAZC&&);\@,H[fXZ*ZrtV P`YTBcZfV( y\ckp9Wp|lXkЀGr}z9\/A|ym :EL/و_2sqh)(3:6?IB Wf CTR+2]lv9>+);=yWBP(J*&Z٩Jen |gx<+=mܰ-r+{tE,,ՐwXNqj_&ԂD*4RنCO]"d+QRn"!K{HJn"ZDf'qa0;?ͻ;nW:,yW{3ڣ.INVu͚,KmnEۚ;ޱ:r/ -zqih}+4۰(-)TYHjjH)g|`Mgk kQ .4Iч@ P%/A]N58 6.U2?udVSUѮJc۲}Q̈3bψmiA 1*9x]zB@E!C#khm6Z+A%ԲT )Df 2JZjTAGb7[ʏ癛:0/Nsu)ٚˋe^=/xųd0ŮNb@+Yh'֍I(I$oĞ;ÇmYǾ|[C>q|1#01vYIp~tI٘d*MGArTc󋤋ZYтF :-$:1=r/,ODL!*HL䱀ڃޕ6r$W)̔A9c FhQ$"[1YE.-SȈ/2p{  fJ5ԃC+rrxd_ ^Nɴes1?+kT`CB ؁B eEA07^ye7Z{tW01sUGZyNKr{}}tYDDxWڮU4\[:mlW)'gSt8+Ϧ0P=ÉJ+:&Jj[mJ4NZVK)D}9ZSk=aվ\JpO ,slXǩxb1c {CJ0B$ecFPn #m#cTJA}Kx>}}ۡ~OvslsTg'mg& 2AAc #\`$F rZG#,0IHO't <(Y TEuQI ikIHERL?+%VxX1ޞ'$ p5%g~(ɻKu0Mݭﭔb{7j^?܎Z^YDi0Oi6 I M rNɀ=D9ʎI$p'x%?߂쭋%qƒWE%^qLf zaRʢ2UE4k{8ݪom՟`Zb] .^ڶv70 l /!=.݇.ƥCւ!*40X}1FӠ*Z7{д<:b~³~~_ dv3Yz !V=%_wPgT֡^NC3c4͙A.g`(#c#b44py:ٝiYLQpżBiu^SNHy5#-=e̵ Sm x$ )[X1Lb&Z ihy5rZ]8U$XfHHF(GfHN 8 pS}W9l*CHt}7+":I)2ܬ\(0 ʔ\&xl|;1>h* IreL!>>ITM_v`ƎpMx xycSE`'Lݯ&;*!5!e'pgUJn 4@'HWY%$4d*J"TGح"FO6%K} N(76ɬ}ȝ< b͛qJu.zymga;vmE6TZnhnyȻ~ey֝RtKXGYzM5on|p9>2o|oR܇vh [v vCw|+68/^&̍r _Ơȵ1O>7&.LvpA*#:TP0vY4a#A2 <+[W9 AL5gT;TNjz.ռ)&x5D%]5/7Mbۏ/"h~1|cg(OmDPqDe!QƜq!s#ϭ< ^KJH7 C+yu=a^UB)#XNqpBL:I#SC CX踲`Ro@7<8_|L%::iT촉 a֩Y" fNxB˽ x-6҄n!ɿd, "Pp$T)'r0ʆH4<M KSo#d@wE_YWQ XyojQ>wW݀ey ?8 6z^f$diIE&/~LAJMR r:DFG)m-H e?9eI|]@fwTjZ=?w2"XA̋FckdN]WxynyF~cꚻaU+JPBrOvUno謹M]m:,LMA6"~7fZ2OGiNnnmE %^"^rvM=Fi )G|1EH.ީ*g|Ӕ3NB4dθPI_}~ÇͦwWxͳ^Lr(95#U55Ǽ_ȋφvYif|LrXAKeN^/& S:p%Wz{?έAɧYPO*+5s(K=R<)P??Yh dφOMFXJ)ܟƔ"s) {PN!QMP\+RWbpREwHJ-'3^{?`V^߉WԃJa&yt8%wa&GCPqvp"CAS;M'Uͮ581f5yztnDSnDO^:U/q%=ZF_>"ʖz]iu4<0smCrz짾۱rxkOC{v̙rʱCGo3 d gȜȥ\KiO49Q)D23R]wmmJsªlMչl>\hfL IqR2eLA%%Q! Ѝ_[↹]v\ռ! 3z8Bp{5Zכ[&GyII?z,XgH%jl̇Ob>dT}8{3^ #y7: ugl덼^.߅j.>`@Itq-;~vېZH`콵uW{^O k{vWmO?Q-}zoH)W\gcA?2zyZyZ,II[l#~MM(o3^Z=LrI>:8e[ d)IeM]UBoc1JS/źL:MނIT}wPuOtIޖ};4{[‚E>l5A,ʙ%eК*)-dQ6 4gp~NmȜʃCr݊WYp߬ygMK9X.yUUwsraR_nv%vo3tZAP_hi0 9IN]D:)#B1`[]6d2S9S|Xߵ47EDd|.$գT V5Vj֝ ։ \큊䳷]U c8ogp?~Oo~s߆Jݶ2d%q$Q0D]򞪃&۔݈JPIK!٦nSp|v4}(˗-T P+B%oxge{>7^'8.ïMzm;k ]G,1{0ޖ ׆k'Z}- hg S.I*gl"Cb'"^k'1MGejCKQlz.2,|F"UU# $\؄fݹa!u {D x~XNn;^Ͽ^.f$6Yi4`Ԑ3:^~*H#,!gx M(5,XZgb-J%t1 m`XBRd+eBUܬ;Kv|]k_{5O"7b@,u2G7]5I$I*u疋a|0m깵_fEiar]#fJmi~` UЖLtϤ=0ݝGaZ:"ۣU[bS 8SPrNS ^I%:Z@Ƥ5:9Qu|&yWC@seaX wS7nmځHw99kVn+Plrv鬥ԽJ$rIRK"\,HLN_-^Vl{.J$0R$ᵯIO5B>2Q-sďE0* 2}ܙ۟׎w1N,et@CcR۔J>TڰxED%#dl K?~j r:"Hg? $"tbqlgթCvK"/[_b~̍rNY:zd(* :* נP-H[c`0k^#Imr5;Mc0ATg'u-mVZ) 8BLRG2vJ2̩ve h!%9Aq{D < *p _ êUVi{7ދtЅF^ :/p 6NxԚ/IMw^ }B /:^˱R98-P\r}uv/Bb$R[ ]}VZ5$I6%K>O09c)ʫK< jPq9VUOD$j* C)zɡN TB8HY@(d',5f\NROSXPIł4gJAdhVJfH*Xo rrGr8}Ҹ6JNEF]AKY+4Zih/e q05c,^lZ'O@~iV<;vSΛH9cRZ758+jh3 l$kCI„!Ws%OuS%򨾴g&1r`tS`T&E" HiȌS$rg9Oc:-^Ozƃ mV~C_Gsbu:cIEQh!zN)q8%-Գ1 q^mCznumN-xagE`̢:HBSAJxWN^N/ׯh0>A<5c^CQ&>+(ORFoDL!r i0 j^)S)X2PьL TF,mCِQB4|-K7H}3?ܑ^X/dG áCu곌nwo3:^jīoo Эu^?u5@d5#xM51 &2Tdc}Rt"hBbWoVea,2UvRkC!!/O7Lqc{ }s7̈>fg Mgs>laΙ߳|5zm,),FW:r;X: LYa;WrUM}Xj=_,:Lؠ~{wOn\lzm5f~#b6zewPTmo;VUۿWuGWkgهZh[z_j!5dʨGl}<,ru?` 5}XL oVxz6_|z26h{1T z e @u2!bA"vӋijziK' XJXWdIYLaZ aH iJ@UF+XIm^]Vy}8^fk(^F:o(҄v^Ce*Gl)|ڤ´Лn\_[cB΀.XR >:eHY2S{U U%'AeLHlCl1@89ZVLV5E*eT M ^3lQA 2Lh31JcT"u! <ͺsÄ+U) ֞;e SZu8!@eeH_JK^*h\"" *ibn+Fp,%lk\ CY_o)d!Y-: j\/e"-,O|TPjur,JQg|rHlawdmh6yoGЫ>pǒZT+߀XRkutYNzt$S` QBhB/DQ bՓSHzPd tlSI@Tkج;kvJ3]le t҅=c[U1bM(*sCtZ?q0P1[ˬYSIA.*S\\Y1R"g6GJ] ? ЈJ6t c;]Ql&JE$kfݹ^c|F{t,Vljhc3Xxd3ED+z):YPccE0c 5 Jҩ>,U+qHO3W+ȋ0cg <[b2ɶY*f7bJԐ|ZȨ )yT5!/YdKQxPUy4a>l |X /LѱMC5̈gĞGbJL;,iB|PĦzW5F>9؉e;=nNpy%ګ<-JAqp>Wͫ$”ჼnRyR&)pLL9'py|'إޱaIgIH':/IF J\`RxitJT@A!T nF+3lF8 >Ckez^>˒ C.qBzuT 0P}r?|1;Mqp9}oC+NvFCn $T/~VSpbVzŠVv/k# 8[Ԍ&iM|A ?-WE+^V#Z< ˫[cy|7YKC^swL N/֌FzA6K| CXMԁoԜmj"M4I!('1; +IbiIpqA87h,:)PQk\$N9s(hH‡c }f?MȑW}?U {K,vVe BkF|7';*I2YY@e^yAjPVB{$قzxs{M=s1#Zwz >છKmkmt߬mq)yH6ɇ ]Y#0T*,"fn0n!><,>V_)- DyOS&.]K |Cԑ}`.BPƹIB81 LgSG'¿LIExj:d~cpޒ ]n0W:O9G&e^zo~fYעKK3IL)K̊qoZmz[&Or 2T*y*Dj'N3^ګ3M"EЙFf4!2Q&DD 21P5x },3DjG#u o>4ZgkS[̙cH"p ( AXN ZZVͰNRbjG_vñV}!)/,˧b<[`cOø]EuS)pɃ0u _a8͇ K/#z/CѨ0|G}]O[5 !Ar68L̿1r͡Q_U_A ~7`y9~$ =G^A KH!)pLL09QȤ#jyYV JPjd%pJ^kfE4 8%}W>wթ Ɇ|~ % 5SFjO@C 6ܘRz}Uߓ-wvlDZ=𧗸G8aUڒ_h$!Vgړqɨ;>DUwNŃv'UL/jTra1@1koX\?pZrmH)Uh_g ;˂_wx=[rݏ7@Dp,MB$=XgnvH7\onQe"gիdN=+e7۳8= zz~9839jkF}8| & ʷs>QQmf8UphE;7@iCp8`w$SkP v?>ݪd70ٻ1SKc.u6 m;˶v6dEfv7N{@T{o܃|͇%Բ%oA˜gٲ`}x=\|uק`nq[ F-9te#M^mnF\q3c1CaI`B g0=W}M V-qӽ_Z V(]yQGiuw7=kt]Q]^\-Ko;ϧ7Äa  HZњO$(GNpTRzɈS 6G>c˧='z"lU#pqP`b/O./͞Jފ>[}^Dl͑'ov|.h2m5^xۖ|oҢZ3JHY|lusٌF}7^RN65;u=iϞMtw[m;^v?D|#2ƛ)xA2Jh`I@KM(\cc6DW̵jWf誢=ԲP´XvpBWS7Sp+k5-- B*\ZM^]l#+ɖgM*\V˩UEibۡ+)s'i z}\ye=] Ci'@Wr纞#Uq]0p0LRp5ҕd]=x18 tUEk骢t3]#]5w5DWuiƯ03`,pk*ZrSjLWGDWdo-;ܜIZVݵV)i9s\jf̮54u(4}49 lE7PV誢uW@)|pcMKcvk*\g[+-WW\%]YQ5DWT;cvmW/1ҕ9]9kS38*^tUQY]};tE{N\ Jx5zVrJ& +깮W)]& ]UFBW@k:]U$g:Bj 0k,p] C{Pf:Z%`EpiV%1vrP woEa=:Uy*Ɣ;|3)@:'rzrs݊_ur%qg=;XS1afwGA[U#Z\V*Z37eo* (3]!]I㘱 s) PCLWGHWCFn% NgI}I݀v[yKEpڡi5\B-M~̮43M#M+kv s- ׺V hwSrjtuk[S2-VN^]Uj#ztç٧k+e97 ?_\֏u+ns_߾} Ys?\KWhڛm@cwTe{B/ήcß:st៾5}.Fě/pnuwh_,oWFK5)}5m{ ٿu7>b+[)֮Om} ?/u}oov=vK܋ukBS׽y aY0!QJCnJ(CH"&)āS1,i{3]JH4RqP:D&qA \QNmh hM>gD] ^!E8q)\} IT&-Lksr|I+JQ* 6E1Ι$ʀ)yِSό'ZS#Bg#162DQJ Pt)RdٺLdt}h`&Gb׷rd Z !qkІ*F $$I$D^ \a4 F[bѫkJtx2"i4Dl^ 3Јr3Iү7\?m2ET--QsvQ.xmc"vN ˡGQXF*'x,+Z ƣ5:*w:;;,T֣## $}[ /=R":Jt`:@K %!B\6HNY(g!X`џrԼeK[aC0n/$y}6{ &'ve)9e9E dGh/&edxAv`L\0o/>1]`"L9&X2A1Tu~A携meޭ84%O!'XU:TMY "a y,'((&'Jz _qXl[jU,*I&3U R9|VE #, #ʄvGwcY娠($T)W P$$_TřG2 K[w *z]rRqx"fH57]6ǐ?31igU VBHP9D&T D q]h"< [[C ,,xD`&f]1E&R& LGvU87d& /Qb%'ڛ(SaΠD]#)#`hT3 o(Ez@_PW ()FHUW: NRR"QUDI)<Ǥ}]d0 ի @^ɑx)%cȲRhjnҰpfcuDՊbs6V뀠 LrOA\֊qsuĬCr|bH.Cw89SB#v5&XƋwyÚKs^'}-UE%vuˡmu`-$8b8 \ȋQ NB-2.GdQu $$BmLBQX|,mt{TdUutѫ RFM PRm֩Q5H  %- HV%@%:(F:.n63*I@$OEVJC2~ȃ آvGyQV"IGvTfmy1 x戳@0c]9]Ԛ %Ee4騖9s4T$52 o^JP)Dz- ;6d B_wG[ps8Y9A07Zˤ"ұ!}HJZ)MdjtN,"^XL-SN٤Ez/g6Quhcj 8G>:58j1/q ;_N=C\[mq¬8';Й658JA=8̔gSu,︣Mطa`?CVXGuSf<7~D.bwح a GW7Ca MZc$<&#u^uoSclGjL9Ǥhc`iF F@α:aVzY|iϤv8  ն>6 6u﬷AY731ZG>Di :ץ~in,U5L*mVaBQSD,;u΀ŬƉjemh_HO)Qq8lP9>Β|LXwv85 gXRJx;IgPn &o1(8Pi?,/QcꦵṶKc"D|0c۴8:6*Z&?35 X؃޸iAD}<4˯1t;EPWa R/*a>[6jS ŞԶrUM_)dmے/89ȥ_5uM-Ri1/0%&?7)ŦnhۏWL'"O}nx?4`2:|CⷅtkL'_,Wxwr6T=M4k.>δZҕV_ϾϠt[͵۶=AK^_cm&o@3C絊qajsrpl@;hM;8"'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qhN 2#pra' D%#ahDN iqH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8^(jX9*q!Ol@QGewf'@Aaf '8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N q 4N#' %G\@ևuN tJ@ctYٳ8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@'8 $N qH@q)ʆ?-ɛ9Rnq_`۾I8_`煜KBp|KqZL7.!%x1.UGI lp"WDmrEI|c+Z\!pF "K+L]z3 .WY }mUm-yeW>Cc^:1+F+:ʣtBʠAjre@I:>rEr+}ꍁ )jrEf){F7j.rF(M\9Kƛ)u0Qu6+_aN'ԙkL6-<_֑)ة(gurXm>ݼ TT*M Ҩ ):?}HwUbPɲwٻ\P;ݬƎ4Yŏ]]a/xUmb87km޿=9aFgjXw^/vꦱ}xK=1o<&vbvPlxv}x{Ӕ Vu5& էт _l my˼Q ?%~Ӻm%V_J/Jo~x}vmjfzS9*bhjN%ct#ƀ^Yu!UݬrJUMNag({m@zioz,[!Tem*W'_U:zֵƹpPnLJ>g2tg dWgqPřMk> kJ(,-u\_Z (+c\Z3TOJ("WD/AZȕ[H8 ul6n6(ŌQ::FrࡀsXpx2GzQeUȐ rخ!E䊀}d#W=Bd}rE"W#+mp\p l q \hMrEN\Pl ϻr\!\!U\եQ r5Br`ǤPO45ԩ8:P3ڝTԜ&<5;3 vI+Ki"ciH.ZH8YF,"X$QZj|riӎ('F"W>!Q:'r5Bj'"\`#WDlrHv5FJ*$W-B\"ZS|vE;l+rTi 8% r.ͣ-gQzUR{̐(rخT)2+փ&E vGi&0+Gw,'+rE^1ʕ5N'F7hE](pV3+Fϣt"JE(WiĞviψX7D9ȇA>YkwbΛj+Xgn'̕x.KVwPd%-~wZUWLt45'3Ԍv;L.}Q[3?cLX\NdÜԗYsWrW/=0T>ګEEs_ br}/;?o~ӊ&?5IPvۆ]C_AߗOLǟ/s&g7߿w>\_;|HlgJndž >Wϛ/0^ޞjgݟϗ3M9~m"),R_: xSjKȯ/,x SA{0l7eQh=Ra1E5𶳩I}QaٲZ^-#_p!<꾁o37|)e{׷=xZQHHHц'zI5^Q꽉> )Ef{cj3A@o &dF޹$dG73Ӈ4w{eh/~ňW>Yg72}%MIo T-H~6}B1{1yIGzxSnTW˫S[WQTR3|Ōt"u櫁I* ^ܑF޽vEOO7_n"ǍUFc G^N1w&i**h|ʏm48Y!AXF &A"Ħi'HDi$HcL R}y͗}G|Ŭw{pt3cǻycGр)N{ڮA^`lw_h!&'v>Ү萊JA1OکARKjq)nNHaRo<-9PiMiݳeGMeG1YfkևH߸d2f}I˩R=lpJ|YV^3l<;+Oo밯o(5޿u-Ҵz7֐6g7%&Zx9I # Hї>"eN3v}:UB,8pufѦˣ J"Wz NHh>rEr+t"EF(W*:!896rZk.rE(A1ʕFH8j6rE)r+5Ɩ.WD٠w6o{ (EdWf(2!N`N|&&P$(L*#"ØE((\tT I0Zm:_\eP"W#.,p"WDb@JdarQSv:D6rfhM(]ҁ*EFk6fD늗+ˑ+FoU3qðfirZ{.rEq3F 䊀=}flb(*S#A>pQ&#rbJ?tgLU+&WyN #Wy0rE<#B`"\z_\e"W#+cU`$Wj#"ܡwhC(]R+ɮ(W+v^+\!-hW\ea \=\9QТDũSOi؃ձ9͎8-!pԎJfB,]2Q1W )Sgȥybq9'줾<_,gܱl`#:jIWQPh(:6Y[FJqf31'ڡj(} zP@QIw\g#WDRr*MNldXrEp+-JQ\P"*峠FQʂ(*ASvQ6rEM >kB+^ȕy`כSÿ'8 І<\HL8 u,2reD{[`d:ѺT\e\Cc#+*p+ugWDY|DEl4#׮po]\eR"W#g=qGd5&hS 8Y-p94&];+2bxJYR3)%՝fgmLϡy+6'e93@9D ;b +`^[lF '>K 沴@6@> * #B`3#S\-~(efr S_ <"H_ξPV"`?$J |5pW$(j~34S"%v)ç֩ *>ҕ- #BV96td:<]EgCA:]!` U lNE>yBAj*c5' φ"\ņ"ZH"JOU'"xW$ v]:5;+et t3]=V8F*FtA\*2u(=dC‚Ҍ ۝'4ZrvWh(CKFt1`CWn\ R+#t$y# w=r%}4?An.WO;Q⿏GSwD~Z5B\ܘb8.(ʼnEfDLvb{Ĭ8h"`hF`UD}t(tt&0+laCW\ тH>(M}+i6tzÅoDRgC *gDWX3J#\f&6uB.-/@՛ڨ 0]4 i(S[j72zm]Il* ]!Z)RgGRFC4;m9:o"]5mjJU~ ~I05`hr So$L4 Z 'A0}'$%[YIQDR|GKN(Zr&k9K(Qf6݂7L %UD 2u(]}+o-8Nt֮lADkdH2U9⋀%W>Z)LW/`a"H=]+ T\f4芆W9 ]AǪ^:-'.aUvZS+D)tt6UB |tQ]#]it*y* ]EFNW>ҕ1z?kHX!c7?{ wz842wcih]QeDkgizh]!` Rۢx.tѺʈ2rdDW["\φl-SgGgG45 ]EΦNWe[ HW!*|֮ ]E&22]m@WvDH8<[u;*uz/ ;ŇnPl Dt({2d`mrQʨdooތG]/Wmy{'ӫK#ZJwpuD`a:UPgf9y<מǩ*'b#A*o gpu3+NE=,l=4Pϗ9Fֿ:Uc`1Nv~^]Gէ墌n{;F_}zTXv5:2eU 'ooS_G.GF7rū3[ ˷uzjp[O;iW9~>|[0wnt땲?G^#2F)xcR\oy =wwyK3[ Gʖ ׵ ЄTCGZ{h2=Dr0:[WyGG_*2Od W\F^a Z-;Y4؅gZW%AmJ>:_fM='T8D_V>TzCljP4O7>|b7ݏڡz:m]^p5uvdϡ Y5l~I%$@X.t䲽Q)^m^ ۇ2.b 4w;Iջjr~,ЏudkTwP'sIB@V)RG E]W%4A6R퍔<(o3=+ FɍٯLu4;t0Q*k$=|ZWPdRޚۓ rux߾[9u#r>?oWrӿG%qn/N7kg'h1X?;ahG{tc8%z?5(x>Ho'[#Pݝnf۶&ju3mX*hҟHVtZvoWnqlѿ PpP7Ǯ2?Y+;oe$wg) AT;t15h7P;EicfƊUkU=^n7ZLT{w=A `eh1Xp ]ZBp;QO$3]T)b+^>A\>tc|nbrq\пKZyY`(Z@YuW h\U ۰;ڸAǜ/nnl5 F"+SW)m3Jނsf{jZO_Nu`)(딞5*=&vͯ'ױd[Eu][-4n%oNK%5;p(/v;lnP3*iP fvRP3-47j5Or7$pGQB.:ِ|7nrn$}7nxjv/ßfqgYg>[|C7jy?72_}7~̼]%ƃCLOA2a1e(Xp-$0\0 z/%2`T4ʚMyN%m1xZf.XBD|H&#IJ:B 6)H&Q0C&6d\ZAIrA r֔DN?O?@cڈ/C H6M0`A@Bf hcq>d[ ygurFHZ>k*h6z϶gylGLф1K]X:Wm*KA^LziN}giSo]Y)#uYVYjHx&rIg;aL_I+9VBVf].N.+RT1M]xM!ʲ jf휿 8Hry)er4r{O{S!jZYY:$QO!^= fư~mbzj׵jH}kcF#q,A5S !Z9#C 5lafʩB,fPӨJBUodǡx>:H,]z2ʯ)rOỏ<߃ h6(uЊM?6"0(ks\օIr)^s1"hdbdbHR8͊!INioL遮EOѡzj 0]&IbCH lR`*!GG2 !C32XOm|wUA$ĉJ,:d&Ls7zm~l9$9ZFZNzaN0cq1GQ&MiSϢf[juUΖR_]<hAZV2Sg%c~6sw``y5G9Z\"N. f k~KS ؛aMSϽҎUItz+lI:fXsUdA^lGS9{~rfsvJRy'cr*YOYy ,)WˆIr)'6s0N>!`:T"񐁤Nc:Y$-ۼ;'3%g!${B 4DoGi-Pexǘ 3=\Cn[ܷkvƭd]*ej𕬅W媦lQbY/>YfژUlr3. 3e\Ni\aB  xMjA2Slg6_" KW)TwX吴Dv1wX m-6I>g`P+̝M<.d $PKC优MAҬqY^ 3_^b{JB|0vT?Vrie 6O~~Q2|MPED;aQcAqЗjR\b)r0rCfbe@3u2i3% 0]Z˜~&6)#WL&H!,4ϕ @#u=PzV23]sw I4a< 1''Yl.Ș&lʤl!?I:?H<$tLa@ }Nh+SDF؁וnz&! :r&m8{읃K$;U:1u")DNNH괹OS%$6؎$yra)X%+W;IZv2 x o]>'R}(:$Õ VIO>9D J92 2+r~Ps,V0#\4ʐPi>OT M6_/rF,mIӄ'2r8c4el!jQD>`n?M܁;$63ŧ2j;Z 9MD`ҧw(褳10lZɁN\:կ#y:琁MMl3( '9są]ss p0 ]FKF:Q,^VC?Uޕ>cE/XUlunoצ3aӖ$}$S/@$R"eRx?xfC HQ2~:>>EzzbӁ(4oP3Ƈ<7R_>)Zn/^j;V:dwh4]jƨ]ѕd+ZoxҚ\LQ.iN~zrY/oMi>ut 7s{o&vfQ@Z ջi<\]8^%gI4'C/lхUj2;Tx{0%5>\4AwO"PJaZ<,sߋrb7J0G7yDO8?#Uש ђa$aAErd@`H/Ͻǿ \Q PLz?#91J?\x+y/QuhQr^p ZsߕYw&BOs SឯR Vhi<U#VYeG[V9*>VL37Ѝ4c7rTaqW2r /3u>aL:—7T֊݊^ZŕSYDT|̞@ sN@{\Aqg߷Ü+$_ɹyYJXu1"@!g, ~i?4tS8lL̇,|[ ǑS ^׫#/4}T%Mgbwnt y9_>-—%˙)Z_ Iv ᳴o^e:yLLnvF2MR?xhzu#2!V듞 J}-&Y7"r4dԚPsQJd(-JZ_^wco֐1>o;++>'+zЫ4`Z}5Z`Saޅ3KY&-]Nl X|w9lsSQ>oTFu1OJSpgi:^m2)Zw^ւr`ȘZ/q=۰+ReV9ev`(|+  渘 /G[e.9&*䔰e6nEΞ2U+vor?tA1 l!rpv1E4dr4Naj_>+{_BJ?0?Z?;j6H:6 ;̼$UHϮ$>^2/̓ɝV|k!n5?u/z&";SUmO/K!O^w C s a$V@#FHa\.cV39v!^ѝx J u`}?7$ҴJ<vp԰9roT.Zv`Dg]aq.mbюrԛ㬋.ՠX?4>0?^W]>o!fN:F {碢<;خa Tߎwl'ƹVr6Hsqo1id.=)d64QnQqz1k9R>tSY[$zna'q*ě…:`S`d^/QSI Ua<9\*ɛ=Ѥ{W?/wtZ7&hT|ĢhzR*s%$6g-&AU3 N7yw7hn~#Qmɿ/ ؀͆&{oj{`o7<9-s6FrL)xa|8q3ƳQ4Խi ̀ yeF#3PÜ k,Qt2jW緥 \=f/8lŋ:L}/P}q^bwZhG !ZU/iϑrw#&AmT0ޫ*8lȂ(_2J6^603hy㶭qvnVmaĨoou(x! C׏Kuȅr+vp=1cGLرjM>U.짋1-":vgJZ *('}Q0BJ )6-tvd`C )̱.ځP' 7D4K _gD1u!wNP- ʞ18&BjUikJXrU @ >g^&_}-Aӽvh^T+GqMm֮|ug?b"٥55tiJ$JK97_gTfz1l80j^fr)u}qoPgNa:,-]7,TSc^] ɀ>U!ޫΑ1ŽcX6 v :OxxAIJ1Q*x֌c^IgRmyʡSO90[ -,Qhe0 L@>a@ >]D+q tNrkm[iVqB[UC*'S$.PHcgVu>Ƃo,-ɸ4oj;ފ 8ʒa{61p笈)hD"6r5c#0Ez1PNS(H:)H͇:SvcavV;=AZpi58Gec@α\JºVpԸ&*6e玴Hm `y((QE?0q Lc_Nﳑ>}0.X؆uiZkQv̾8~^ j颉\KB0Kb3ְ@18 :3aLL5`X Ǡ$'" ,UW>qŒXUN\|f5ӈFb*,&HJ׵e)6%zok?"dEKc_t`!]F{-T:˛S1NBZ 77!S# `WAJq*x (~rtd"a7bDE )봩$8mVnӂlɆ8*aB\ʠ>p$fڂy6<>h@6I+C:]9[ )֤fqg~HM^EWAF3<./ f0Həv۸ V2"#ҁlTЌ ! 26HWHDR(ftuqR<sų}4đ&Ճp2M021a`V6A4ޅ@T}'>4X-j!g4qLk"LE0'd5 ##_z0G.~Tg1 8LF#9/%tx_=@"fj`4~@AG&ܾjF-]y ݾ}L2M۾}P:`4K(p0;kǀOOC!L_@/ 7RxUV'Plh!␑.鈥<kY`a5jGYw!<~ @g:t`]vw.hcF"!& QH"HcDϦ0%O'cVj&dgښTǑ_9./8;eUQP@uWʀ idc禴YyS日d";Cˁpgpcy-rV1 (R<˄0~($U5p 잋9匿RN t<=N[%T{qx%^rd#K^)&4Eoѫn(QZ˽%hT5ƽ$|,I\@&+ֲz*d|,v|iK==]}5$Dv9aҾ3ġ%Z}F]P`i[qg-L )ASn~6Npl04ixy ZN.tlΊ\rU,e&r%>da*s/A^q4Hi>C'ѥ#ǫ|)}9W5H-hN' B`dIEn¦ù L8oO+=^ȫ7w<W1W:R-<ʤ &G!Q*h oCHc1͒28F]V=:A=  4]5_mDncqOKæ"D؄ᱥ ᛦdZv/Ux=/"Rܗ,u_!g],[NkMF< ae@[( ޽3 ,hS!WXcsh$n \@*%$R1(tnc%\rЇ=-t+k}r0:KL9G& 關_p0- c&eHHPO&x}DT /4pFŃ!)G1-U[QQiqIҒ?yK%Bi`/kv![#OٸVr畉Ju#nEƇÙd:u![;Q@DdyޣӺdFpk}1(ЍLA,tuPTLbs]OȤUY#XKN]kL+db~; Ƨ A2xXaȧ8u`B{ o2}u"%cxe,^15DʱHᄨ8HàZʚ֡X&ޓ4 *% $ܰ̇}\UP>EjG4y"lqU {?EfCg+d )4A!t24L>U.?ty߹u8xNPfr)JY-Yc8+Y1kD-Ɛ֠3sO źj3\RQ.?x?@M+d@)NJ13Xk.!0 9< ".G2uH B ;.HuޭEVnQM.|K92Fu{٢-7T0ݜ%jJ,bnQ/U&yG'si3±g*0>YBWǀ2'WQ/2yGꅮ1!.|.q UEP_* :+WHl׿a.jY#"拷ix9O(-´gmÈ^v(cvZc7BA8`v5zx$(6W"k)C[#|QnN^ ;i9σoo*{jzPdDsMa%88~|X,n$V$UN2,"̓$E riI^CBgtPVD̼>_ύ\g9tX$l܁*.ې=Ǩiy@L8 º RqC`D0kn3 2- f@mWi$$ЃVȸ_"4.2K.rja°KI7|\\/%]vZHt΅ "˩1`"uIQ]{܊D0&?6, $Yf9 *#]9Z}|~H:qs$DvnW77Z!7Z;3A@M]g KP\ 2+>/FKB;T,g]e Pfdx2 -UuUȸ.%.Ou@Y!عP~9cutoW+π.$ Êl| 4Z&h5gӆs`9*i@0F^o(K7nhyEd7ZF2I"iRYQadV5˕#" K@sK dZ ŤS@) A@e](A͉@"R]-6c (n+YXQ] _\-Zຸй`Nj@LrqPBHCb3k͉R%TtWVV/w-v׿vxVF](%Fuf t].B K5%VeCKf>QVg!**a@#(r|Di7]j * fɥn. ][R*.X1OkT?Bh+d R0hRj{}'F2Jw2{uą5LttqƏq00faQPTk6!|@pXL\s9ύQApHa",AdaC!F7o%eۙ:+N*Y;rpꈍS+}}\Ӽ2 \ 4ejmwH!A[]&Ɠ?k;ɩVuJr$+dܜ3y!IFk Ҹd6Tmy1[#E-!,J'ǹHQL?gFߏKItl[3wK8ḹ5;b:QnK/ @[ŅK%(øÿ)ϝevE o{es`Ÿ ٙjr yt 7o6?I#WBbkfsW|\A}(4~'7K"[S*8Qgw?;.iC1~f;Y6E ) QӬn>W"WZϙM m-E7nD 3xvr} L;=f8Rǘs,*_`CqV:gBFNأ!O\.p~ҝʿWP7$&3.A.֣|Œoh3gJ wA#լ† ۟2HxcTYM,)vc=39FSswM8ٯe2uv%4Y_Wͯ45f&vTFIlxCn{'yOZ~qOQMaC^! bJzBke0gPFz|MG2T6IOz[*HQVvo.ܣt6hA ,#z! AyNK!CiEq("7MiI<<R4@~ȝR)ő[O2LWVgLzJh6ôKWZUdmt A5YqL8Rw[g{ ^7{?HyjD5N7iM)C$ᩖR{xyӨg7Z6eІK˪ ?ǐʆH9\ƽ'c倫H6EBiK+shB4eh2l:5$/?m4\ǼJxsg,UbkC<%\tK8DF>YH BvӨ6hMW`a'Fy91D$d2+p%W(^+w&aj+`*Hxt)I׎Z6xE:Oqn,p(m8^74xyzg\3RµV ݈r rK;*ڹ-g&nC˼l>Ts`9BߘOGS45j{ op}p{fٝTSm\8iTy/*4{(~Fw;iԣ&qySmRW $s %4OiWbM6s5YLP;Xl[q%;3P3ޝ5ZBN=z 'F⌿BցXWBHN T7-q;եT{q́$|ؖz`B`{U 2N =VmzdM'>uf8]Q{T5JxJaIₖ,g`qXX˸ As֩/j8>uQP /}@^[ɺrMHӦGYRfA;6%>d1tPїaiw14޸B`$o&geRơD}z*D[\x,PB|L㸵s\L`= KC & 7,G_9|@\)* 0 EaTO\?9rw'`(8u2ĕw-EDfj 8Ke "J(hW^7ںRP! 8i-SdkrBf[jy}Ir#Iŭ0p.Z˙YkGϺ缵LٜZvZר5.ib!HH[!NjBp<\oP}dY)P'5)Oޤ\q+O>nanE"G"N rIe6DJV@!4E DZ(LUv$P{m;*ZBqrEEub*Tz^,d*0PXa5q%Uur8qtWHt^ktT!殑4]IFBƽre_Q{Z3*p1]eVHU!&1mjM^upއaG $zrTMkgİ 0ǰW)_[R *d My ܚӽ s%3y5SS %[AQ)~D˽iVj2t\pb%β#ӁHJnANҖɱ*HRU٣KufӺe231mnpVV-8v׿v8Z:ooTP!7 c82j *2 {ƍ,``Iuݤwl&|X HڲeI-G'W:,J"ȒԲm;;n̴9H:%Z\~J{i[wW`ߊȧMA'& _IG]4O0~,u)Z l%I֟vlo tڱaaAUu<`gOѠoMF5x^.~J#=^M'@Kp},߻Iӣ r_A6%ƩYRHnAcG/q.OҎcA-mduܾdiT U2xy6(GW s -kgA$K}!ҹ^z+Z Sv(p9{s]Sk\J:ٿ -tn  D5Sw!aî#YĶ6ߴPC*ps RÌ[Μɮ 犧 !(2/y÷('s~nc9Zy=Tбsf=}.V(_Y](nFEiqyvjs7W}`wʜB3Fǝ;tQNBJ rOD)} 'd"\}T}h\5DJ.*80};"u9H΄bRSժPD#x>J~fs^vqocbD_y%͗}@&*5e֛ K$rK.٬;w&/A6~D}"O ̄q/zؕ!(W=o.h$Gs.Y|ѻJÇUZHtyBgs2flyYuX'M snلRݹ[f+= 4er9&]Q2f7hE"bZj1uP7y꼿r((h"grmͳlx1禥!XKg2 Pҽe~d\KfoZ`% >@wd haUg6HI,({&T*#3VwsM/%ܵ0#\*^QirEG',Zrh.tK޽fT).CBZPKb\ Bo3tx~mSAI颥&$P [ģ *( 4KL)&4\VcVHcQNAX sUK44,NF6;6@^Mfi0$I81)J&im`4Wf ] ByI@jř &|e|GLgvxbDVr9o;>+!V(T\c: \)cؽ]IJ3w'@jW aa^D,ص;F~-N5*ƩXb2v0<޾ =ToIS/VKLMFb? %UEߤ0J%]uZ@ q#} >! ڍ7`iH첹Y yh_~il/M  ڭ~nnj!TA{j(+aQZ 0CM00p,,O"fkmǥe&c+Ѥ0_x|`{oT5+h`^ܺ=)"Eeͪ)E:5kZҴ,VǞoEQ XhbӮ-#O:{n-Yֈ yPTg_ F.1Bh)iK^1bpTƓh/!,A_?|}[UWzK3ȦgJ™8A)CZDITDna?? 4=Nl3hps'5J l͖FJT:7EqW Nvu*uKb GYpJ8Θ: 1I e9v*mMb5Ū0% Uên*Ou7,a뤤Rcu ֍)qn(-gGE:ş,aJ(Y6R&J򜘕v+oCʊ~{h@2rC = p}}}Ƶ@apCܳկ{E>=g}PܦGVX vzCˏp[O̫(SA}.~ TyO%3ȿ_̏b1;N:G1ST^q `=| p\i?%3#q2M{iNԒVZUmtZ!/-EKVB.y㗻UߥV$Ng7p _=?~U,̽ 쳒6ymvh%n9_nO8j*cz^RcB1icw ?g}"XpVdq{H2fP <ijX(R e"3BDAUB#+tcGJVeGi֝By1i{ s`S2|X.*O~!\ ^DZbc5P`^1l24);ͩ=jE5LjÈ0^&7c7W?]$WuZpԄMp+9WM;t 6v:=-ɻ EF |2HKb<Tn/J څi;ed;-}r؆R7`Nx̯f(y\/2fsP%jh @P)TުDoU]ONlx0? d$F&cSɢ86 =ht?}^G_]kQ[ŭK6?mx"U2=;im"虗LNZH 9lXw)TK.!2dk@vGa4/ze cTpIK<[Ӑ+ADlF>z~5yUjLBHeZJL\J&YǪ8:JJՎV[pkl S\*2w.#>Ӓg3?pے*6$kgGYZ>gu x j`rL`Qd&/mxI&!0ܧ``]WK[]&UW}şĒr~c/ 7-drѨ``Ϩƻf"nr6z-rWVHRH0TḚMٱ0[CK\ j9EFMoIPDFʐX*TcQH4W<1+.biRmCQ7oBmA(ISk}Б`W~|I`p&:|v.\LrVYy8mhDa+oiz*zRESJdcl~JTaJb'_I'Bh0*ـd;b Ȋݩ[Ov@<1}GD,w91ʬ*uIOaiQRk 2"JYI D6h|wauxUaݸ-Fv:UnJ 9<6)#!_?EN$RKĹQD,Md11ͺ4n>ieV#S|nUw;] 8'͌<Y  LK2E@50^[N= X<Z"J9]"uL~0B>v__dc@uOix}mޥTE0O80G/;Rh؋S|1K;so4L'0Hk2rF3`d" HPN=C4J%IM4~X ̕ a]<`OWj:kTc4Lo@A >QXdZv (eھIj;+Ā^v8SR➰t0ZdWRatb\} -< lǻ =s100z&IMd'C# 9{mUH+v|Z+b2UFv%ؤ.+nGy1fϟO}J̃hsqtcBJaG aZu}T\DpW7Uych7k~kW9wwOFn'~e0k6A9uO`Fq??|p/Sn?w_v߸LkyKsKH7џHe˰A.Q o >ҡ.2s*?jR27/L * 13Ġi02= $ЃyT<U0T%iPL3!cʼn4V8~ ק{}^>_ X(QYI% ٚD ~N3KxsbpiOSƿ_Z]-SD@I-ާ"k㘻fBI/B)EC9~q'{4ܱxU^U>WUsUIn-ɸFieY A1DF"Xc3C 5 ɔ7wԊ;]K,_&征:c3@d84C1͒ ,8 cVHgXJm\P BrAkJ BfBsd"iN6U j$e[ǘayaXpdn#=Oʩ>жc\!0m5բI 6f()#8d D2|(p lۛו!Ue%j IÈ'䃾Fun h zmpZMm@h&q)5c8UB1ld<,12vY$FqLp lQxM<~ pWyW}>jnA*iir;Z q hX[ĈEQ8U%X1cQs+fp+Ũؕ.Zc~ǣ,tylT ja%2PPX* #jeyW~zZLZm2D%we7پzpb^JÝGBBZ!X9 ^chl]^q܋.cGO Q'q,uǒaM8J%.CYhʌPI)uB-Ϩ]A*̃wx ƊYf@\+K2eD$v}HQ \Td (6qd9cfwHKp`l[cY%S-2uٔ,`03"몮ZZ"6wC)BߡxGEvN#a.D?J1[*AeQdLBct>l@Udƻb[F nn#ݹ)oR!!kiS:%8Zi˪Έ2#[C~6Zal=5mY4jM!q]H8=d+&iꡭQ$l [ֵ:#DmJsV8-l}]-ӅtI8BLF a,vS0ҥVh#j : (pRhYJQ9vNV$$yq DZ|5G[.0%ѫY^W i) pF-DF|Lt;wjR2@޾f.|&P~GLJJMǧo&܊L0u e5Ȳ` C*lږF8o uB8T{Zw%4iqƌ{`E%Ʀ]8jEɐ[iִ?ik$'R(kXɜ*uڲph9ZXG99Y[b4৑pbpڑ~fwx1h1x[!3,X4Pf{O38)wK 4w aF߀oJsd9+ -JN?6_' {lO[rr?y6a0J)Z-R)z7 ]WI*0iR Lj,6Kynycxw.Ht[+G^i5x ܕ)0JvtqHwޠ+3e僧5^B[gm@$52>`CXM%8•b5S2 _?*~fo\PZ4ZXb$ #MB/.-ZYXdgl8t5J FI8jMvbozg4onLέca{%E``d뮥(2jd>RPiЦ^ܙ! hN jgCE\@[N\֟-TD[f[tsv3"knIӗj' žx(P9:g Y[19%5nw{VR0V(8D-Yߦdzgsl''A #H#K 0130K頲\K?j|̽- `Th0-˗FOOΙp6e7 Z~-˜5;i-d줓\Ai ab \סփ.Ͻ=gݕiR@P1Z'`Gj<8D=)/}@VtXfI^?5vU4$רJuVxJ#-C{BFyj=Ɣv{zRҶzfVTve^g[g|g; ԤƟsk /, n3 {yK>-ʧy=`2?Og٢y#r?SGk:!ՠKg%a=Z΃|ϐoy^'S6XwIpFYG 勷/ϢD g Arc/4C2X9eDge4gVQp' ;5Y ~r]/ XڭFBl:Nם t'//Pmh wKFܣ_c=ߖgQvv`i7Hw fL;]Vd<]pb$ڡߤKIn} eN~嚾s|VL=;yԳ?^e EZjJϑtd\N~ǫb,gc߻?CZ|]9q?ˇNqF`Ec\?av?F_<_s5+$}* S*r{S~-`|8d=:װ*IJҊ k;e6/fCV}e~f/W J4lO9mi~҆HtQQIb)} LE8>dP5c Yrq ۉ;cꭌN c)IQ@ [7huPDt+`#IYH=0%Ar4~tЗJ4cf;3}L|P{MmIV;_DJ4$W9at3 C]pW}]'*ZERt[ɰ~DeN9u ^EG*Vudb> D'ˆL#ʵ'TL~WPGǙѩ|nY^rqW4_1"2N:T#c\zM=U{dhޢ4pfLFSR,i@`FKRI4?cǛۗy[͹(?2NdE&R)}xf2ک2uTGs= S41eYGub_ gkp |晣Dc9iYO9c9\{K@ z\c(FnLUQhI \V,k3BﺒjS4xfQpZk,6aGI8cUK+/at3gᾨ CDv ֞{KL+'R3gMXuSEIW%oIR4aii\J?1?Z^PHU#"+M{,_Ȏ3du3!`;b~ٽbn|mb1̇bTdܡhP'Ԓ/P/96MYn]]omJ +(5'RZ7ŎE;뽕հA01 cpDח]k0B}f  &щ4Kf 62vW܏c3.1;fFh)HhYFIP( .nPF1k-qu($1qUb6Q5H%B&f1BƗ,-,*!\M`;}yy o" WD9eE52VWpS0H;Pީ@j:R bg0;Kg`W۬DP ?2`d)Sؚ] q@JK1L9[ Rx_*%we(GShg]%khByq=ۏ?q/}s;A[OO孀KŸ0WY㸴6Ck-6i$֞| uzOMtÛ͇_AFu[4Sa(_m4a#|y_$^J{;u>sgV}]ܾ !#kd}a3 ?@lE - ܾ6t]#,}g!f`jbiZw79* Hb\ ;̿4Ĺ)" N}"Uw) Swީ3:$G!oS,v)92~H$]. u}ȧ wmRsAR.pc)e<Jq̈u=v51c"S̾5(bqc!=<#nZ#LZJp؄kd\01ɏJOl1\6?\?\MO0oe;|fRYa Nc[Ba gξK=EQ!AugDA!$1%ֻ(t c=Jxo}ĘՎm+S%/C .ǨV'FȱQ*:>ii|>0!2r+C(.ZkZGjears0Kd{r@ynD&Q%FDžX0lTy6P.ϸʼsqL{_.Abg521zJayH竎3W^ t:t sW{=|@Sra!J78"ﭯ w_E~޿7l :!37r!jgoMD- GsX:^d<9q?}{6_ 5qV0k|q9puDQ<-[\/$WMϟ&Wg`i;n۩l.]sW ~"0Wd6/vvu QhjPwP-\d=I"~~> j>bXrwxB7 UU zآ hP) m 79/JyBhbiY1ĤC0i иq^Ԋx+EEh^~ YUNi($ RN rKW# |\ jL7DkbD-\D!bZhjdјB\IE[9%JW8 eݸF r ^ny# Af kHto0>P &9gmPW[⺴uH}J%7h<G6MVLKR.UPqEcJNmrGc0i!kvP '&UGNLC,= 0oև;*IB&NwGvz>a!SXA͸#RxeqWTqւ~+ՀnNV7R! !!A{ GMl]3sNKAzZϾ4dH~O9~*^1zP֙ M#pY|îݫܧw0p&+[LeN;EۨMvİ M?`77C1PEg{v%&o!F+D'1\p͘)v.ːw';nڝ\ ˴f{L=z^$IvEu=&Z $NVhd^`R@cJ?q>w\ 1&(P<ɖ8Eb+u &к0Q04:x4Θ6~}MsdͲM{dZk 7<D"ܣ@D?I`h[A񎹰:dS(RV^;ebw’ʊ  zz}?4-CʀF.L+r-!pbQo9yšdcvϷ/=)&԰F޲ƎBWuc$ظPRnfw 9K ŘoFH.#EB.$\#$]Lh*{f TX*9E>qrʒSiA ۧ+O?oxb9=#l*9NA}̩ LinY_.IRk;\Ll`??ň Yd΢ Xc?KqQWLtWūzX6U,l_.E_#2\>׍KPxˏ5uYob6S&Jri'"7fy7u~r%f:~O)o޹1'ߗ%ҩM`9x;V#|* ,FBHp.w # |3B=xPDl܃ה}OᲞDi׫+:_|of݉mGKXZpaZa/Z@3 Қz<`Y"{Z4,nY0쮰w4La/$ ޗ),Tz;bm m S|~ ˹S !2`R eƲ,f8'7|sjw:7}Őoǵ umzU %ܒy"Vlѿ⦒?-C|qD|Q{:z*=11LL ÐBakB޾ hB0Bbz3_-MngZAۘ9OfOnzG s^V8vGuxrT^ 0xA27e?nTGMFD;t2븧prx:oC޶:3S<+āPNƱw l2S@nG٥B[p3סּrTd̥nOL~Jnj7zT x|zMd ^a@fIiJ9vL>ho8o>'Nc\iR|EWrmƥlY~!/\ܯVu qɡTir2Eo"} Q,P{PV6(s_buo6C4UUMͥQۨ^8RhszW "6_M lj"ZYyR R>B9+4+j7!7 kiEFIH<211iIdYE(٬YMDEX]D &) SNpdQk`DHd}0$k(97NM4v+hݜtY#s#u2Ԋi,] fDbjl fv%edda]sHob7_|7v9 q~4FWϟ=svSxydgs/'˻'<oN:kNo++e9&cf;1sx/u9X \ZjsxQSo^~ؕ7йoӫw{ȣMaj('9Oc[os2qn”T*$Q A0Pj`d tLB28R@.-7jW!Ry"%C3D_r[S4_ e 1p4VK99#瀡K=5qKѝ=׿a *l$L {lI'$Ұ8N+ڋXjW(=H>f84Glm$R-9d'~ #{+8 AJFR)kMMfؙPmzhiX{Է y lYjk>:_p ʚzFie2&:-#J]#-f0E ԨFVɴ90X'[WIW8^([݊(wR^d0:boh(Oo{yo81La༓LqӇ}+TCI{hK5ShM/\1>ƍ-Ǯ Pi34*("r-5I1KFRlLxD$&]^`2aFfR QB85v%ș{FqtΪ]%m"t48;k y_?<̏lԵQ,m4ɮe"BLArĒ3TDC6cs)w qovKR{)F1xo @CSq%㔈S FZ[#8 |OwG,Vԇ:F"NKvd]RSiKpBBM?߶4b*&:1r OM(K=`ѯ5L:pd+JlpD{{@ ffrJ.`K|#WLk,~~ap+H.L@]+DSd~U .* {| S -8L:%n qw4I) %Xe AGF>)j0v(XGѷƙQ3L3n.!. l`AZM~C/uQXAB.$;%HkhiatPj#/v]N4]Nrʯm yn 4y cv:-2EfgV5u#15@Oݫ=jȁp=DDb0fL~U_VDgNTs0(FyNuQ%N|*1VD[j܊ X7)=ȜAWr.J w!Oh];x'8sgթKpؤ[dӔDGڧԛqXr,Qƴz 5'LT+jF5$5, sgiVS;T|a8XֶZ!hGz!]*N~+tܨbiTqwlUqYw8m^ց)yj LjS$J쵹q4sI,:$/Nhkz@TPXg0`IkB`N(,_홎OWΚn.x/K).Pzo) : y)ŋ.E0}0v㧋,O)'/vC嗋Y$[d̪YbH\3OKlݵj1J3J73?۝ ZC!6r$82єTSj-/*JWLB.hFז}\4H-|K>;Or\pA:q`%fM1Di r.wr o|9W|(XsZU5W"f2"Lg9<:APsN(ke:b}LZ"K4'ʙn@Bgo+swoX2qJ qM2F--> fBȂ3p3g™mҖ☵o 1R7|0b+|w nFe.2IJԺhbCj*ڇ^㐟H-XCghYca/r(yEzdD)Hmi-t.1ƹT^ Yq5ہor8uۭJhΔ=ad@v=m`ξ [^d˸{<^Co#znt 7}Ns]9nxy̓9g)>8sFgD=$oS5ߧb#ds\nUO<ӈ 3 p]qFÐY AY +uc38}o 4{fMntξnkR[FyLIf#"ĭMJo3}HuYaϾqxGϠ c[fP]^`fPRB3GYNW'E ٥឵.0:6CFAFf |1ɤJ7ir b»5n~}ưчW&љmF60g7zsǪC33[C.y0!1ܡKPdrTy%ɛnpV;inSp!*newdPSèdJ3h; b[R!hh҆&|nqfΣsF۪)0톸nІScڲ;sÀL#xs6xQޠ펛>DAo!ʾ'ڠN*FȪa;A9FΡ-c]7]3 7hA[LG7qk`~50U OӃwCp{n|/H}!/_IȜ0;l'+]Ks+'f;!2SR&w5DM܅%bvmU))R* m%†Ư@BTMK-\Z "'ν@8jK,˿f$6*΍p3>a+ss[?)TpA>e2DjG`ƖwɱYtCT-IUZ9a6^ovH[:!=Z"X6*F^!1Ӭ)*+*I.FKCᄕ1ylk-ĒFrE %%EL\=B#AamKŎDBeG%į#4߈njt|R0̰xE1֚7?|U _4jc̳=TQ ISӺ PH's$U/-2eS}bAmj 2怌K%K[r. 8V&YZm.,ŪX&dϾ%Aj>U#T'[u5y*VKlށDd0f=GZ2AcJˋŀa0W+?#atGUKA$HrJU'NHBw0äz.za%.YYLsJr،[%ruTѫʆSU"e 6: P$2 Xr9>1 EHb]pT qT{X'~P^ס<|99͖K)@ Jj&j+ u duj5e`Gq4ajh ( ]i"yI?06\qs2tLY U(myhf D~NΩr+}.bսݺ%VΗF;at=j̽H'n]f譱Rg4)QS*=|YN5 eD_'^(-Ww3ËAcC k@qH^O˯`ymҐ(FEq8'Mrh)I!f5 h֜@OJ)!C$d[J,-Q9/"o{:xXOէLmMmLrj*&j%)ʱDqCsEd z<<θ&!~D$$8D$M@fmrg-#13s;ߙ<)Mp#;_]3T.E`d0 )57KJYzn a{UTO:^F_^'2lr.|mVxT8ڷ[,jȁe_*/Jrg۠OqT:3$}I"1 9m,㲪>z`;bl5Uw!tLǦg (s|Zm*G)zWBFԚ. Ft(ITvٶe1/MOI3;l3;)=p*k8T('~vW}L,$5fןt*jLa')T҈3G ~Uzb_U1E]) 5vȷDuT96Z藨p xpčkv6lHޑ 89Xw+ɩ9W~À7xB->u|[Y~DRu.#/#NOU1sb 塆} _P\JQN[17o+P挩.9y#jN5:!G$k<"=eB?/^|<g..Ҽ;?[F"bQxԣQiVc(6eR3K9+cCjSE\C)]֛%F"R Cv} W~} f3 6XH!G$GBJrwG \+xXȢt=#Jѳo>.ˣ̮dcˈ f"Ʌޛ y֚4>zZgPF1 t r=%G<͏@5KC֢c0y!D}o;gc\՘a#?aI7Fʀ y AVw/%.W(֟N.RT}uPŖ;I޷ɝԛ}'M`;{!Lsz_b?\Ek=c곣rm4j-qPYj4z=h*m^Za2:n%;H 2"ظYvd߆<7`TY;a !Wq)#GCeIE9ߤK ';?I޾x}tuws~}~4N>){&6)aG¢8Xr! VQ(Zq#ecX{k!-qt(ս6~|%h<`858CslvC33DhुBf\ zP!>;9e)u,8YpМ,ztk`;}h7\ϛsAv2['X㭯MF #*1:muDX )+Pk? ծ GV& 3/) ޞfQlWbk jg^Fat E#K*D6@UYr6YmPɆ+vZ$KG&6'k^IWVf)NՇ)Z>1NIĴˣ2T&F|x+4F8~tjTVqa}kF'Tcvs:YصgjQ#H-BɪCfVͦ椵h 5?]Β/Qm:pFz|ww3ݗ7pgb2pUbW>e᎑aO6pGd/gGmK5ٱbBr$,L ܙTKd-O4~T/X9T>.8U{8HԆGMiZZ&T%;TqNտGGʗUX|P,/z>+^ucygHW|'2?EͿ#|;f̳y7XBuz<>Vshn 7ϱd_;yȾܓݧ*Lfx[yPCQ۶qk/eNn2}SF\SF>^J}gWQQa)#kȔ~ď ^ܧ{ٲ|p'/nVX'ظvżP͘gʨ3%EVΈĜ¸ ^Pꃞbz*oC[zb7vGyrh#yt`%YZ#~9m"ȫmGG-"([#/DUG{ kQCz1!ʀj#my&iNM[tn IN̉V1 vS}T-̈́[c9ZgxJIGkh7@į4􇤡SWAF]'=Xi/GC>ENB t9"߷\n[E%>!Xnv ]t֮'v(;f;:Y-;WhB02nj'eBQ/g3Âr^?W>AՈ+Qy!kC5#W_Ӆ1śΗN&]}(P,TF*7ot7wWW`jZ})mKtƻfԙ7Y*^lz$Ѻ8o<`ّcdG?=@݅ RĻŔ VzңiY3pG?'=v*Q:/0:σEYJ>dD /s nS֍2;iW͍>EĎVhBi+顝B+5;aZ\Vhw@p6? YԜSDȻj#BM뷀av?9e%h:O|&uȇkB/xƫ(Ncb .Tra(lWd")&o -Cͯ"ws&goKCJaϣx6uZ]*u2FlZlLu>Ub$PBS]xD%.)6:J C%&i% wF!SWTp }GQ 934E\ E`Y4wz)rR!y|՟zj~tdEe^ AŜ"2!V\7i:_3tE$UT%eteVceT3>tjP0QǦFAZceFǍl:8YSrlj!{U%"*/jӵhplcj"rzX VPW O4DedTamUz1o oo"»5|[]bT"vlJ ٟk ҁbpжeK\uA U2eסj-݄2ɹTՇ /u@,]l.}:V4 -TעMQW,:X1BSZC;QֺR? !vdMIP]:[v2I@UmP5 Ege<.O^gJ |?dVUf9jЙʕHk< e =8?4|۶ K>xQVBF[to)Hd+aWșUdžhRg`qe/7 rq5P(~W`x-։B#XK (P7$(rpqh?ƥLoj0Kk4+= @lN#mzu=q۶rVpZc] rjNBRLgM}f|a2=YPgJѺZ3fsHYg]o(D ʹ{s\3zA\ш?S 7.9*r.x΍m1-Q N۶~B;n]8fp{';p)ˊ g28Ѣh$ N\}a 3- h2;ǥ- z hI@ݒQ͙dAЖ\,`Q@8@YiE%}-4?by=be9<؎%8&Y#hF4Jhh4"ocuɬE`WY3/Ǻ5+s1 ~ {IfD 'mMSRٖ)s ⏯s̋M⑚LʾRv2 5IfL\ѥl~58ʚy9<3skas k%Ͽ̞'ތ2.}6X@iG@3/&PI}̆$}*̌)"VkZ.n\vەn'Iu)o55~vd01غwCCO1 :6n-4{Юz?x &1ag\le̕GPFX $l&r6"LX1u.޹/cGj}aD}gw8TzdpE Pwj Evy"w 6yY^w+Ly°bĖ\<!tňEr|d6Y1Le_eT_NoNeJ=*ϛ"%N]x۞5N|[^iSoɚjCqu=xO+5麟e1n.{njW:d -!vb6u#[K Iط?{R; k20 1f+ݞS;Qc콠}OGdK )[}|.vFRaMjT ́Kc$vJ.v{NDpHN0],8&Nr#pݕS֧А}*0C[eiCަ+NofSm{B;d.fEJ5ރM lP]u>rޗ` evaW7'}wvŗ` kz\"`3Ɋ wԗ[2{ˈ!2rg5A}UG͖0 gJ?Q+fb"^-!٧79ۏǷC|㿿>~u쪲D߶EHlP7]c]Ysؽ<ŞVӐ߼.o4M5th2>4rot˧ppB VG}DՓ+Kvfý{XQ|ڡ -gWl|coǕa:[&]b-ҔeҦ\9= sZ72G#g\c 744~|]%Ƕ;K< y ˏjC+WL[9Z\$?7oNKOVKj1 Fnf RJ ^ y[ Mhդ Tg!Ї'FL豪"JFR;\&`g\ꬊl}{@!Q$ΰn}\Z0,TfYTj\ vo+΅TQ,̑H%X='GIlXR g*Z{JU8WGUdX=Mv1=Oȃ;er"Q|E(jdg*FSMEq&jsv9|C}/=XCM\{NO4]-_WC`w*jDs@H:@L" ^[)_]}AUQGV,9@F1`6WFZ}Q6!Q;'+<0ARh~ծP;QGb0Jf`΋TgB"Dst"-7fvN;!CA54jnE4.(j'WM,4>pjv8Pjn VE{jтCvS1QLB+ }t"Ylњ6Q;T7y!3E}!{Hn1ݤvCppPd>̋Jڭ>N|v;B,@zwPdsvQ)=sɊq#n(RPbn)y<>0 ߎ;zOLʕ]P`10kZJV@SnG#9s>V:cN]"}gvL}lBh?B'B>G}qv7IӝXul`gL]}pu|;w>W 7*Op'קo_7xdFgl_߿~}|t}nŎ.O'2_=~(޷7g)9ʟOouф߻7ǿ> Nvy *_Muҏ9BܒkT$(¶YOEK>ㅨ{>u]),GĆ{EH)(G,CyMXK}X^&TA^f9s@}a[%V9SV[fLaے:0L-hh0,8B<$WT8Mzf9Dꊪ+ϸW~$1FJh?/? R +EްL_KxdmFzӉ hiLe)ۦ7׫!Ek LK1b l=)3Tpu claTm EFa [6%9)ƘX]H|iXY|ccv:VB\uHU1Y8!:c?NpҋZ b-,[E?/M?|| Qhy{HŚ)~< 9r' X?rs\v׀5Z"wgpQFY&1,ĐW?| 7rNwҁwv҇ϔǣ`Dm;$wZ8e9Ɛ Rwz5q/(Ew%p7{<[o8z_x趜xN3L][9+/{vwԭ#aو=qÎP$7n\flo?4 EIL}L3mVWNvA#?Rf;S¨ bBq5:o]HD7 7_ jR J>5(^SyEs%FRw<'䘵iM.uٺQT!ކvpty1N cD1C$&]ˣJ|cT`Ŭr׷pZ"GIPsOBa hOV ꌳSLS[NᓎŢ)fOWg^!3ǶJ=ٝr}>CQ`k(%*;!Ӥ[t.ncgʵ2;;S]U6:*I5q Ofx% Ÿ䩅^PÏ\YF [պE-=֤RG`:<&kIܒ D?;Zcz*1W tqms;m%`k@(G}֘D sx@6mA|q۵; pk"H3y[LM_qn^|6Bw{d:^K)c8QK4}+5vhB sq@D1̽ji <ϳR D:#eTk1c J-Fm.*IbQɓe50A]oّh0PcL#)2] (9^ٴ~d\~fI?A-PJ6$D m~r-+Q?o1+]UlȉQ Se!,5_ /C'eľ ؗS0{A>3rE{ǜ79gZѐ e-晒^jmrꄤ>j>btO+t Lj5HOF-y _h cq"e[xzZ__Wi+BT=,Qshj:(Z{c꼽OGIm0sJ:*1c"l-^&mZ.^ *lf i*lYA">)(i^Oxc 1]Qjp?R^څ5/։:Z'VK]X "2o2*or&susf,P PjEgv]XMjj:̍W !BQJEfaxYI^ t1+M#ɷ^'+u@$'6*YAmhM^Y&hj9j&/Bp 1LdܘN̍ܘN̍scN[f% Aq4g 3PL;62() ~KӁ9 4Lf_?b;ӾUC; !I;ءL2IDUv1A jTj|ڝj3`\j'aLTz']R힀>4v';$I[9TZsh)`*vI7MMݙvFE%v V `2[H]R6C1J?څ0ZmO5%{~`x|)#Db7FLs]MsRSL bȶ%4}6fa-x>뼣J;]'9Q#y^TF?{fۇz^y)np=^I_~\g`ukʋOg{]/'>_U5HĬ]+t eTz9;|KK7i>ni{73DX~jk= ~uZzO oКǮ&d`v>q{9`ŌPBN+B4<]%ShX)@hꖟ::-Mls:;T3G2闂ӥK@njGgNQ;hѷ5iv8%A 2 yȚ,KaauiO\QĐNzc$`pe$ L+!w ,f=#fc4=DŠx0%ʜ } ~0j?n*rۧOcþ=5+ l྽7#t;?~z]#Ј+n}է֎8aqV '~oofi$(ǺJ pQ~՞og{N 5`(3E~[&ޏ`W𴷕i)ϵKNգ;HuV(TaTu+S2K>jxc;sguPfr~w.#碬5 nF*W%3qQbM`_TSqyMql$zhCԡ"EA/*D8ŵn߉]Ѵeڣ LFѓjBF]ʈlyNN^Ǹu\[5+#3.Ӓ[f6&Z(}hO[ie̵ ~>7XQl ~4#?z7dP*߆FK(g9X@<"CK%ˀ+ TaW 2"4Z%br V'I.[ѰWpCf1K:&+ @R[gt}.MF[D~qAht;w5#( ug8\wSS Ohh4)#~xp~D۱F7nbįubNl)7+e H0S;ƌyx2x0 Eip(DgbXȤ'hF$@I?7o)K!q{YnvmT;&tJ>WV/ocv | 1z^ZrvKXs'DnDV';u?ANi =fM=*e0+ȖbէJG짼q&Z;sk 9$kÊX1***YkZ{8/7*i>K^Ղ4&=;J>!0ܷYAՀݾnW@w~O+ ۇJ:]% {ۛ7B2-2 WIFiY*GTx԰8`їyٴR &Nsv! d(#2"e҇@x AN!l;~ 7!ٜMܙ15ߝiBz~ݠy(dV`X{SܣNe`M/ڵOgo |]$}`TbΫx6EԗˊZB!*sm2i< i JBkb\\H)YNǚzl2LIE<1y&λ<4Y{A?7(x7+?$P@ȷ$բsטгnr9˭slsIҘq&H/ `d{"FdQˢZ&' g>rb2cDCTLkC됓z!3BBX[ ʰ"c$*2qH9f¼ZJ8ulP~\mw .#g= |p@`q&9:8@&u8ѐ5pw\$<y1e3~q0Z9LYAFA9H-4`Pl#phWjURk2/ pk&. LogYF9&j=x3>',%pvڄ\$F8Bs:?D/c<6Lh'JХ/xT@&^E& R)ģ 8ٻ6$W=*2Cn`=c醐WtS17x/H@X$"22"Rg 6 NYKmu,wƳ?lߺyhAD=dE~%B9$Rd2LY*d?] {𩺇-;ELFeClmYn$l;YlU^ÖhAGeY@N( @e%ԉ8V]Hg0c!FR ؗυZ<V,$vhTm eJ% t>9loGS<[d(>g~E ȣ˨~65bO3|@Rk'Jx6 xm0`mxLC Zb(ǴO_VqxEk=E9!H, [,S&f L)pS> \Z/$抉\!J /o˹%D[?ټ+L+jƊ~`+EE@~bE%зB6g,tKpO~y5*1-tθ)8`$:hb3qЂiiq *s{ȇQ7EI?̰zL7{v)̤ b]@y4Z"6 ώO!$ϩ6ncۙ$T| D)-rSc>(zMlPZN_.^ˋfJ?_?u" S \.E\Lp[B|$#\m`l&1@ \sg Z^7V;'dX;ݠW6hzQ_pQ epL70cbkڌOD%2iI3 pZ`'*zB0VIJ+0¨Och=C4w.Egb qL )f'ڐA܌MmdJ PSߣRG`vցS-~ƍ6A__@7Un_Z[lrɣ?=\ӜS'zg}=zR$t䏬PP3.5n`1(!QAù%Ӑ-=yq"a,|F{(Hu>98j k~V[`KO?E&Dh~ x~}&ՏM\f@h;Vn'M-d͋ێ?;9'p sN9vpΤ&WJQ XEeN3D(e J+ ,̕&Z^1_.LiXُ)=8SU GgM @;+ Nel{(g)pE6x=7*;3ΔJ+fbBr 0,SJ[΍Њ[5, 9 *F8U3)JGS0\j1Eja'2gNkz j-h4,maQ^4q$"ICTKgi\"vS:HfDfCsNTϭJgt+eZΩSKA*B Ϣi="&@ > n~0_!!U ]ғQM`'{a*W)JsN+%,TA'!x(xB56 B퇝&폝*e-sQ%vqQ2I\DIB&S[SPޥRӁM*D]}_ /ԫխYW)PyA/0+F'$r~5~vEI˗1 ~ǰX L#$ )xJ7^ B8++oSQTtq)‹Lpe#DȠ"ugq\&'e"1.0 O!d'T=qåHRLJ)2|)#X'F\A`d1_t.Fpc52~ A9{0\vM* 8l->=\vJ8V WL&jNJ%N-oM<7SrԹ zZ`ip8" CH3}E+=]o_QK)U Vh˧A I~,L"ڊrOċ$H{)PGLsg7D]M*$y&`DXX;O"Y0܀AP`Z0V}(Rxzaw <+mG*OLߝ4yF)hluQ2>kLT.xF6hTJMYutzS:*JAmTİ2#\KќC6(_33Q*ݴD(VyWh(Qf IЇY?HuOBOR-2`>Lb؞h;EIM7wvFģЊxhO*Ne7!A1tQ mQ &,X9NRYSOv *pւ@6/; @MǯiיDCU[IbthgDΔ3"xk#) 1DYp }wP\xp97>:</Kbᦿc"Fg09^єe۷=׋3/7l#o\y0q# ,KkL{!r R&hCAvM$4AB$@Z$UĎv:C5YC,m 5TpI-<1 gP,^}"q%pGphиOemʤȥR<շ `x;Fu>úZgXgwtc ЌbDtB!I9OeC8OOH&z:aSY^yE@xE =߽gԃ_o"%ޝd25ذ$ga,e𤳌:|Q|>˨GZauP GGq:bQqM9=)Et96+uP: z37˵Lqoxm-9j-uYU:T79*}vRzߟ}>7C g&$'y 7cP0A>6wAJy?jXj@bcL%0U9( !1hy8Lڄ 棛NE)dau5N%('NCx GhC)pUchcr.Yb倪*UYŵ/9{UƎ*|J*W$+nc0:lVSBZS! sx[Ֆ'u!LS%\2Tn p¿ei?<7x_O\j?n sg8ml܊l-_aܹ`ẏ֋٪a}wy4gͣ_Ȼ'<\ɽqMS{R! L̄~wy^!}"*NKȸ~0}Le7.-×"& x֏@L/_0w:n~tl)3~Z|?L:!Lu z``8:%~u'3O쯷t&5{/~{7/{{_:~z9Ii)z J.H7G2WeqMLLKOs 6o>3_> v~n& _ʿ́2 ?^x=S!s/7_^ { no2/>͛`rG?= cn5W^HW~ lM/`w;`pOyl8|6ʟ 2W{ӟ4Ƨ0᧷*AGsu@ ^>_fz*EBo.K̸mv˒*^Xu9eڥNZQ [5dq9_B(D >K CAAͥ"]zkpg XNXKr< rK?q+zO~:)'XΑ{1|)]>TQxʻ26<2Rr]C\O gk0#.ijҕ3FlwȀE.,sOjʽwEQ%&mzľo`C,Rh™B 3O>7ԙ4Lmw9j%,d$򘂺4tD qlJmb% #5w$a{b&95qsD[-jT7**O2Շ~%X%XBjnL?*=Ôv.5w@Ku.515ƠI'6d2U$4M:y&)FRwm!"; 9F\_Ѕ.b77+oOퟦO5- ObI龳f1+5|xUҖ[O( 8q,4؋+m2ӶQojxC ?Z VzCE>ꝣ5iq^ Qu9 8߭i*s8浙Z EV؜*#{*G6j M"Fy1t^ Sx)3NzaÜM|d#e\kSD\Αc$!)Gyx+6ıT>rf'[Tb"D;m9T6Ӕh7/$;Nm9ÌTl CS: *Jes[ C$_cs+0h|]=h%ZJD@={WV>; VSƥR&M!=m<áϛz9{;1W?,Oߊxk2sp$f I,J(Q+K~wa|ш$ .!~Mzf< wkIS }Ѳ Ry2'\4a;%$J0#(N6㺐pT[QnYNk°_jՂD@,Q?}-cC9[CUs* UNPm+JCF8Vٞ@ř -zJzWik?TYg`DvDmOm.gw |z+yB>̰xt'X>j-٤rx$.'L:yП7aD!ψ44Qhλ<7?ھ嫂-)o7[uΥ\H''L 6qX1#<e;{ ʘg3Y ѫ˕l*;DodQk_L0bCE3vxFujz{ T`w">o3: YVMqI7uK16(eT" `2 3h0-f]X[EP]hKy5!Y.[E{5`jGƯKɴ_- K:'0;S%[䕠9~M~4imb!괨a-Ңi-% gDn(PR!0PݾQ ǥ D= ݮA9in54-m g(F(Mb8Z 4 K^r]ۜ,j~*>vZ 7eL@,aD;j,Պ Q ˨iReSi5s/RaιI ӌyBm'ꨗ}O\v가 !fL>*ao)x~y~qtǺ}fp|Ћ~zvL |\`b7"E ǹb8 hB"(cEKRdPR0"f6C4>CLZ X{4TyHHKT؆HqԮ-f&nI fe ADƃ nƍ9XL/LOrt> :qE RYO5 |[l/xd*Z@xrM!Ly 0 *pR}ZK%-/3r0 o0t:W(/m:oNJhfb//wþPֿdl<L{5a~Q2{sxj|b'$Qփcx;9k6r2`tL֐2 xYH% i2%L=v JM2.1K8yTeQkL^_c]A\ x:0TjvX#Z~}Q~Y"+D[4uX_{?m)BޢwӐ(^%9  $`f&MzmC2!q2!qQhLo`GAByL&B.VbT(luBI wȸys򵖻׋WA.A!.zr;փoI2+=4I}W[R0Zn%Yy<Ǿ{c9֋c9֋rt h*wފLc= ̌V cBiJlfzI9xz]K"Anp#C~"&՜:E"jO 1p7a e00 y#/.c@ƫ{x,hk4o5ʸHO]SKzReuF;`UBwK넥RSK e*I NySRnghkZ\`a =JA"IQ [ғT323 Xbƅ4#$lHZ4ud%Ob!k[m 껟1ML"`>ͬO8*U=w[ne̥Wa|X!삣A|<{21ó#HILQzj~~daGp'[ΕO~`777p}xwLhvc!FWq=V"t #QҦŬt։9X* %1RKÏJw1P.Fkz,y0~B7*X!œUP!>VI7ÂDZ I4r7{bG!!w iFEQMREpϯ2AAesTg3:'G9'g! &>nrCܡq-Nۭ8ki,Q'֊ ҘS"Xcx #CX2 .Y0aRdC+۝kS l/}(ٍ ,L ac/C6 \S/To!DMoI1LiDxRF₁ǩ"~]!FV8&Y枰tXZC :[nmr(~~œyaQH +G >3XyTcB4bt;c"yP@Lf>u{{Yn*"y.8vfbNj^ ~qwP}?1233+WQ?0xs㌢:P*g(IE \DLHH$1$rJͽD܆|"zLRg9l٧M=PTA*N;#FD Z֥6isOIiBT-tor|4$9:#&lla"iƇ" %H- @~Z㭷+lXl}銮iu."9bBbAaG;yhuob|jG@b.0#2V;#.u{j80Z. 2!|;t"˰>cu5FA4?s*.;y#e܁}"s!YΙ7|p:2~;??8_*iOZ~Y#Pۭr:&QY_ҁ֕ך=8.%#G号rd7&_gĨ&b ̢ (mu:Geci{Jqi`Gs#ę 9<+N3 eB\3}ড় |Z d=fIE/p?T" *NZ8!\٠g+#ÈX8Ol9;|nZCr<%v;ĹD* Gўkﹰj-VTqz)lDez@ 7`H[ƎxS:P@D %y@R< )qJ<'Mh l{{ZsSްoy׺eR9C> ׉qd|i.*@S$VXG$EeZ8X6P0rxK\4y-Z2&_BD=dӼ5Wo1'_bl<@{gӹqwԜ$Y\6VyOmԣ'ͧ =GM\:3/h~_/A\n7@rQH8Mto"d3;.VL!MhF7T(rfmL-}{.az⛬q@=יXg1k.3ȘY@h~aPo/*\ZCOk\tρ$N,v/jf'= pMÎK͊b}doW<&?? ̂I4s}LοE HB/Ie5HȹT-9YA@hUh^=XG/Nt@ϹJ^K¿ǿ"wޠQPtV,_̧~^E*W'*d AHUJ+c`BI덒X98&ST1gc7ތ/B]ܶ6Jt74bZ/b)/V/bƗ<YO}Kz?7sjLVZ"ڢ FT9Ӛ:.4B\ueo5$|Os:4rAIf`5Q'tXk$zy^H·$0{d4_&X|Waj]qbYӨХ$"k ; 67oCJ1?sPBI(ޜ{U隍#X+<׼p^د -yQ'iM>~A&׳O`O9+t?<'G⾧ZWu Эy<1MAL kw%Ȍ: >)$;k`{*|^.5Rx1  9GPC <()Z % +̍q-P,)ЖcRrhkUYw\!y*6_8xݾrSvmٜ 4aL kjM0 DÑpOmQb3? ,ј q6M,rQǃ>\E}p᪬: Vibδ9d1u11OY.vFy)MhP:-pQJV$Ɣۺ/ect!G$Bͽ9IU2X,3rxBp;)-|)l9Bfmݠ,n}7l2*Mr02BK>Zጱ+Jd@4`a  $ߒIαl &.G 'ϥWn2ʍP;*V+DLzH^)-jr'JEZPoW :$~zbw(Ѩ1)_$PhVd e"&oe &9 FRyE$!)ʬ~N\G)ނs`ɉUzu32TNe\,_Y$ĬJQMLjɁ_ %4a}dtvI15ٞ5$)(EM߮Cr7ƮH%qn?.42;hҲ,XS2>`2V̺.8V>|lX{b{l~2tnqp1LxZ~w.%r9ɝqTIq@xs54&f35H{p82NrcBpBbe8o@@8nv- :ǤwvMV1/J ڼ?fY="v]ugQXK_Tܳ{Uw*{<Ҽ:(WL*[ #&0BRB6 DQIm5'ufY3VZl3:P?EhBAueZq\ U5rVǜb\m8EAhc"&c2LQ!3jLYEŰD8MIX< sc,*^i3r%&1u8Eh~6od/_DN!15+7MlH8Z?.;7rÒZY"Y$hw%25{ f4 aF7H9ϐ:%\W$)v icj6#nExrȰ@$cQd:`T/2ˑe Ib,0)=D{rS3FQ-Ju$F~`vI/?IEhwHSqµjE 2 @t Us*LVn8R4_cr%ZBSR#0E*pێq"O+bD[R )J$w9 S4VcM.^ϕPfw7 2F RH)RX1\3ۓp\`?X}35eçe@Il?3=~N)hFc>C bTzDiM cubt0q-aL=b{gbch:_vqH1 S͹т(!SP3E^.M$G98#N;@=L(E>'KN;&s|l()a^OR\(fz J*.Z%@^ ɐ?1=Lr]3Lmz}\HS`zQK؄Wz>[_ҡy'^oCwKϚΟy+ Q/?ѻAkt\3J͌S2O8(b5h۬)kŵ5s2\8UdK8d1~Aԕ>ݗjLijmL ct`SB~cru|=4Ȃ5ց *(ŅHQjI+ dEG4\2sl.pgu9P(x7v,PѼ(v9P5]1d>n?J ' JncJɽĒc^9qFA8/>Y+\~?OOtgtw7<)^ˇE۳Dcz?'i;qܟ H Kje7oK{uvn[}`?.f rGO\#kW2kH^<pdrTkFX6Z ++ O,MTM.lWH9WPH`@۹kR'Q! \LӺk hTxp 9C8G>Bbk%WUDՊ dL{$aܛ?F1dתTtMW{蟑E_]{{nJtf$:JQ^l{!HLUF_ZJ9<+eR]N:@>~AiIi#Nz[yW tH+2j qD-ym)):RpD%GÅP—'"tUB[nx_V)aY,%KqqJKQ,aj #jNșMF5{iLq1b^pC7G56=xj+_{XVp:nmޤ)ߤZГk aNӚ~js«Kd;gOuwǫ3Ɓ0nO50ch.B{?eHx7hz%T%6#aYpҧou\ȊzfHWP#fFȡ wNٓW3O|dҳL Ωh-3)8'/$ 9Z&,B~ʴ +,}J$K0(|ZaPI!ajv)J_X 0êoV P^b\XB2irkqNk% @+VEPsjfeUINZ,Py"wiVzL#)aŸLAWt.C4Fb%OXNT4Lyc*vPS1mrqO,}-q@\keW-Kq)W.,G8N^@ŽTTT5ATYZ2UQeqL*mС%+|Ԫv@2 !bwQ RE0FOysTVyMA ުV"C%)ڑRaW~)8K՗npA/겺ؽ"ݻK-|)s§\]Wl8{hm8kYo,#* J^I++}B V#?M]w/Xc-W6R Iy7IW]f$I{S\i1cզW tן=.KD_54$+mS>Ua̯fʸJGTl=I{__WCI|BWw9B_B`|@zd%)-Z>i4ь)^yXL8hL~c9:GtREYm] qW5AkYiE9(u<ٕJȪ( \  <#- {,S4 9S?Pfup멵T:F*U[ Uu%RU OZEJv&SEXl 8{j2bI"%8y䵉Kڷ0y2#V4)3ʠ[_{m{B+1_}4QzN$}Cyt[dh?=nTG"{F=#RzUǀXXFf7B"ˑ~HKy7 Yurۋ A,uU;@bB%LU 疳  jlFwu0 ϑFbvXHhNgKRF)! -A!<)?ݠ2ۋɪbXpT>VJBv5q '9#9UxԨVM$f^!;# E-ASS'Z$+*+zU(z]Z5!l_Nլ4hRp4,OX&ARD'Ľ~,.?L^0iCs *VrcM%U\V"@<&Z%h*2ʨViLWQ ˂@&ܪ;mU !LXQfQJyVY_0=3JYdF^Ss!8Ds 0e4,`zPxB 14U_k#^0:kB"RRAhډ 8K(aUNrY a18ZHb9Ւ '详KYV嵐R`a>$,87H2J!('2^ FVK;5@ӾwH/7?Dj=ہt9J<Bq-)EgC麗nN;rr+TLX,hwBq-)gFEޭ)}Gw[p:5ybASMDP5oε_2]j$S+%L9%Lv+*y7* f~*ƈ꫏'UCnb `>KBhA̙$*X´]RRb ӵi_ԟ-9SP._SxZRQRwia?sRVLɘfDj6ΰqUӉɂ[W*cQ\Ij`Vk.>BTƠAƒw޵>m$w[uvqmJ5 ,)ÉE/ @JeY_?==jK3IU(Fu8NUfXKLENL'$`0Fs tϦ]SfI :"*Z#Z5GT HRf>'ӷyv3~V9WI,%4&n&9'9a 3r)t^l>nmEGCAh>s@`v]B [*I]a A CB"Fôzlq׃T1CWRq۷c(zjV2[LS+x(ەtSg'WsϚ9OFb/j ϙѲ #ĸ-hyj$K;C{΂n/KJl3Y\sjٶ4P&x:.͕ؼg3f9%q[[n[U%K$97 ~wrSEuT0U[b\`W}m4RvfֲuBHR}"T'7!͞vSnlmaV0UH΍;7[Ws=>ahqwz{'ATIђQ^˨4ZV`抚,zU(<z(xGmh ߤ}̑RQ&(\BZ]o9^Wf e%_N~͒W(EW*ɐzXAc5eS1Q-uXg.0M\9HVa=#Gz͚?Ѻ:djsMUZ!1IMD ©R\RYkÙV8- Ds!d;[6@R[{pEr4ܾ~ցNyK E2$$Ab I$쀄萐<*R i E}*vS}*(h֌^Hv CX8 U4; UXXG6㟊 y6,g,gLvu>yK XJ/N!Bբ L4TFr09gڪ75Ñur#X@j!Y>zc}@1EۯS3Zh P ]]TR~fSx>0K#KFqVUfI&JF}tO5!>t)D1te?>(ډW:k[j! *_kU>gk;[`Og'^;~S*Q)+#!đTyŵčR[ƇH{+PM>2C䈶DU S*[w_~#A%?߂) ޔjH)Oa(F2K1n&1Kvh;M![*%.J3$T"!%ʐT+S2Js 9\;XIjMDM,b!QX3d0.lRvbz:KW M{X a AV 1vT tBlOHtUr^K:q1$Gl۵|D|v'xD7brmHZW.=GYIvgxXK4%X'\CP T<^)y&zuI׀ab %n2Zb?O7j/vf_8%Sצ-KZ&9"2RBލwX D 5a!,.b: koB`jr ق éG沠P+ʯ2`uY|m@B$kˢk(HZ=m+T<6)0>]1?o QfiS:BA3uff~&G֨:dfduu_x>{!_.S׻L]2u+vEX9(%64eH5Kb*l5* :`gN8 < 3猚9&=y5aU;PO;  s0I壁,3wA;{pfSylD1˂,˂rPQu0$idbbbsrFIhTKE䄲؀ Uu߁Dw PE]u$ Q+"i.%"k;p&g>&ZItG37T:Ff1Rmx$Zs_ +$Di36}Ʃ g$-^7IS6ı69;&x)& *ȪIe*MBl똄Z|=Wzd/*U*86[NyUJOtt{- z= ~qWqͿ-g`wF .p6]sE^}O.T14*+?sgB({PB0;ftb=|jb0hHlgR9bjkNW0eŚr7akPCպQO $uPXjz|PY5Za4F渶Tw-ꆗt<=[d8(0<IF>g`eKق_!\y2\\B,V;MPb l[Vcmg'hԻq@<"O_Sb,iFrd% Fk]0tU:'>T=nG&Q8ᾦ`M b)ƒ jlL̊DX41O$UYӮDKl)MXM؞2ér=g لsMcb).pZSYeAJL z1'!eߕ=ywY4KrN8Rp ̊riR0itJ gҦ:K)kBH?/OܢR jX2c9II T$'5X2nRaPZj6g*1 zDhoĥ-q:jvI:m1PS(~N@Sw`|Ͼ %5A:'Җ3#}fTj9S8gCپ*ڧJH< ` ڣ?9s~4EYMG锃7lDy4^B0(1_e=4:c3f%gpVj^7z_47m%;Tvԅa=JΕU?UhXe4+3sjiG.5 Q^ TZ$w :w/)G$OKelVW. oWJpB2 VRp0p RH6Qdzi9F$>~j X#*ƩY F6ViҘ# *lnɜ!JeFFHTSFZPdLH -`ǀE,܂z1nD!yK#]_Xi}_MŃyL*OzųUnkKP]vO&/@g$fi&;.\B:'مftlxa`utfvܡus:;[ĥ)'֢Iv'A LDMPvfn')*V8x!Nݤ=cZy%yh"/O\g4M`&],KDt1.0)6qm_gdr(%^BG-sJ\m3ͪU z0~UHzp2 zwB>D%.a9O%gbiSJS ʌ|bxS`>E & W4|:@wkJrsQw L.~gn!`>eN>uL 387n~pG0y)տ8;73χu($HŊ m;}s͘:_maߋ~YÏG?޸кY%{d^o|[zI}Lbn6tlw)#~6SR%?[c|Od+ || BAtT,ɏ&37N)D G˩eRA~ge n*͸Jb,s'lw?:f;%nփ0}%p[ jpHWQ3R*Wu _Mf(z:M0WEݴ|mp=5k䟽?T~bɐﳌĺΦ#6y)_xK5*Kr0Ǫpbk;ʈ|THTo:^U"y0vxꇧWϚB4f$Vf^-y*7?93y9m0#~ܤY.Zoo ]y=I+qbiG~v .s]_Q,W얭v" V}bX,.nGw{Ľz1(_/!N#h\*eሳ[l&ȫ.;_#b>z꛶V^>sv6ށbmIުu]OIIE֌> gaSbLI7\{QsNdzh$ 0Q Dr9t0ȴlxJ(x$* Ľ9Os#+irFr$ Fw={L#%%I. k#Jr%iHg'(%.^G< I}'SrQ:Q2t8hy57BaJqNx|&+=b^Kđ 18h@$(TgM1#̃!qvprRzAUטo6{ͰJɏr7ti0]Gq ȭRvz:80x5ߋIf~wؗɔF*>h7\0NAj4З~fGlޛp 6pk+:R̍F2릣lrN[=r4FC~pM)_c˺ DX[*uT'&֭C8l-YкhNjBhO$֍UthR1:m4n3[u)4W$mD/|T NjM[(TYdAZSh):}s˺IRKC떊AImu0n V%۶nɂֵnА\Etl[7UYthR1:m4n]oI% Z׺BC~p)^NRBUn\Jqr)R^]JjZ35D*,6z9õkfn2]qpQG֍YPVq]E-kMWN)]rAj?R @ ʁ RO!*uy 4߫S8"Dj(H XՎָah:^XYu&>0yt?m8tvtg/`1^w(Ay7g3~uE4;M¨bM͍1ƨZ8HЀǗSOzou$Tnԏ䢈h~ʫO#NCEg57:"c-q˳_@(X>SsXO׷n0~SGyCKEggưp{bm0*EoOE}N=`ջ1<ś7_Fn<*94Wjb:}%ލ"/4}8c]qvEk-g\ o \ r l#|pK ~A Ʈճ5P@Z W{\:$^:_Ъ>$k[E~%F%d, 7[8]8Bdծe(#x(6ĸ6%tHyCF9RTVꦆT .N#jFEĈT5ZT,\l̯0H2)`0fF;n2tm߳՟9WJiY6yV\;r!hPVf&P%%7ኄ19k!`brfaN {bD`dn 8'6PBⷅ0?XZT.Ǯ̅55~(2BvUz`Qc*3fl*04²Rj4.iZ@\(Y>3l E4ʎݲq|59QKz&Jr: ]JW+J.tXmnL`)NO[; 1k%~) na9 1RσZ)Fn\(ߙ+BŚWW`HR_[ҨR* ]8ЯIuOeWtF>WeM$lk: *">KjL[~2Q͗dήwOզz|%0w٩e2kVYGkQ|PiP|Vxo{a.f-zA3p0ƊO\.Î&{M-;EYo-/[-RU?wSeE5$3}=r7-\<=)Ih7ʆW.1T"\?mv`xq;+5L!Oa7xRN쓽f(;ӷV·q:^M`^pt8z䋶 @l9#M `0{Յ r%O9OxGTNͨVmFz4ѻ_ЇQgOA剕w%X5+} /秖 Wjw)w:%yW45[5ʏh~N^\ѕDGSl=J|rzx;s%=;zw~C-+&`٣-cG {=eʹ/Vݯ{йktuk kk:3Xbq@Ý>d=R:[P~X$V U_~{m՗̬VcD-P `wzzUe8f(ݢuqM[g>*F %iT'O֛u/տտŐ͎W=(Ti$U{*R)Auo c#TQ&W^JXF%sU*Z9R>M>^d`#Jg8Hj *)& '|1&9E )GPa0ʝ*(XB lD 8 Υc alD(B>Ataǀ#`-`,7`l`r-(?`d"|\G[pheBX7{>cH\N19ĥ$pA/_! kZqS*acSk AOiCBb`#\c1, gd Ć`WWqfWb68A\u#qH%uE5fk-kEřHYH0`Ӆh*j(KI,e*߲V5/hv*+V;Ҵ/|%,'d"|2Ã3_p^e2}fԅsc̱l>畳u,ojCC~dn&ͮk5~\ܹ:Gfhi MaedA'oz1{%Jqt1wg'R mF;.kG0B Ϫ0zHeu'}J/<.W_}XF|!hP 1lЀmĈP{ŔZp,tbKh<ĹHJj-vzhpױ-@';Id➼DM1Ú ENH.rdpb;6WJV!ix!/(cd6W%C8I)FPDY'&խ$7(~g֤F@ΉU#i@ lw#f g68h>ܛSlRd6@Juڴ~n;+泝}bSkk֩N*pq7Ohߙ ũqfs#d6fg%RU^3O_>#b^R oabf˱ >/,}QMRyMf="Wu9X(`5JkA FoDkDeF+8= dlw4B@)֙ "F`B\`X$|kӜwiJ$b{uPR1+F_^݇E{V5RnUfHJ{f׊ht$"% ,NYy!nrçHxGg6 H*zި&rƿTr?|ve"w:iD.eoAtb&^}P2 T7){)6X U̒@a,Sp'HQYd3EK0VogOSgX}1=DN}#UW41@R`a+)g3 U"GUN0GwKs b+(&N;qkf!qN_mR2 hh,5G}yuh b|"!-X&E0:Jer27ؘX ݠSaF!I٪vvIJ=YKo6KNdN:{-P ѡ4a<Ձl,9U COzk1MV%Vt~he0{^i.ڼ^.hFokߌ~-ǐX2d 'Jb%vV=ҹ6!l)* HviHoP.YۗʔA1pЂJ m,eĖpŮW^@.ї[5x>w vB$n[b7+`(_(\Pu Db VFVdȼ9Q0nux{uvS@50` }g6"LOeWj@0H 5mo9> hgwϝro J53˿1? L1H#:wT=9Qo| 4I&-ɫl$J`45Xbl?nrW`blvvpw OPLAup:[[F9fNl>%fu)ZR=6t'%iQ4{r_EDJڋ||!x4}|h@JxZA[i$0Ȏxw WHpL,tz2Sem @)6*0N ߏT^suIeij75v -њR~i7Ä[8x 958y1(-5b0B;@-$@xA/͆=I+ 9٥b0J7c?ɣ`ѯXm+` c{;`wSv DgI廩:Ee*0 5B"ڱFĶfTK´K 2p)&=8h̒ ki(_~IZgYʡYu1ZM#y%-%4Q,5:M ׀"gNnP 'vE`;\1 IW#S: =!JA9`_ϣ$I3>Jάܢz9"i7SS#tY5:Yϑ\"d}N2KyqJ_ˬ)*#af̳(jVǚ[)4)O~!FDEsyt-S*a5W~w#!Z`~"S tC xd}QU5aƁ;/ԍ(೭>5GEKBY P:Kݓ'cf^cC L=(U`c!0a#k Yz.hI8%M)!Ʃ gf;RXˆ l MI>SqY:PsjBvnC-,$ TH.o-P@d;n6XP@G+z_ӆD_F] )T~6& :x(-MQu]o̾zſPoa)xk@S&"n6Ro]%o""[ jhC[/vuoFШyc ѹt56)1;+[pk|YTtP?a?j{т&]'bxeO[)DqR gv0Tdy{Հ#@O˴ο{x?7GR7Xv@=CÊ{@@k3JɁY.-猲{2l?1 3 ޟo *u<7[ΤVP6lx_5PY`ۍJIű_trXc؉<怒~rj7][ ؞ԵuZ"s D%6\( F1 ѻ'Asbߍ9Ɗ !ޟxD8dÞ6UM󨁆$zM E~5ǀHAwK 9F)-ѣs5Fy-D}4M6F^u9` Aѣ7zS"8-D V"WŘL^*BXB.DYuZ:T YQ}5-KcHxLϳ ypwYJ]{rkkM7JYʑV3_oՒ. .]>Ӄ`gNyfr+fZ3M Y=̗|ZZkPN3nC${ɇ8:zzb@j a8$-|^D?Cu:ٿoͿ$]9VY,ZiK`C@7L=\q,XhopJi^;% B3 nߥ QY??~3P4ْݫ_p}Rx{IÿUzzM3cM?#" >cL^86N-8FSU:1O^)T_]"!j(K'Fyj1]b*]vafOJs`!RÏz}%Dk-~ pw@$}JF8aq `8btMي89 虥ů&U0%%TY%C-a#:愧R$Tjp vXiAqBuQ0 = x! ],a<v舵f2ŷFZ/ϊ-xHte;Ll,;xu1 .?yt4*`6 w'N!\m23/A=Tzi~T}vI~@/GPO˄Eg7~S`0ꊮcr Ҙ, Vrj'-=06vV ɂ@-g~v> fgB([Y9 BA?D>=[S N*~A nw!Rw0:ín;w2N~z(ݍʼn(= Ci̟;^L[,I fK;G8 !L|ayvHP(^VfAQw^w"DADP^1|Y# '5{5 s 14dylS++s6glZcT *b7/hV1M5h1 s&V _zآ:qjWe.y*^K9w`AC~໾s[?dW(駏*կ= gOJRAzy}c{$h #y˜]ZwIN#RGt\$'h\- |]ҎMIԻ$th!GwGF9F}]T|u3>cYg]}]P|qtlAnRNo!]Ի$&'%*&cߕc|]RVqrrS;a:dz`d t"ŧ1uo+Sr/~>χTН*,Y<~KQ<8Px|n[UxeЫv 9#rqw' }H&6*㟺sWWH&ri?6i}sc"-PusX1[;0Q:,-T> st%8AV9ٜ4lSPN A{#RNbCKdTlshu{NCNH.oA0 ΖY%!`s!9b`lF,E>9 <ȅuyY!:i]RrރrICƄԝ-Vi)6-A.(FewLCw9 rf ^28:AsԼԄnt@JK[w anMh),1O!!beϝ|ab!F9u;m؄FR!m)X肼Ռ Ȟ;u5{]ڑrȼnV[n?"ɐemsl{FVD^6jsz$尰hk0hC7Xl9y|́D@T1yu(i^ql},v$&"os0㏿t;Ep V鷗KM$ȳ6U4 ܸLg9z>Gm@s\kIA7߯ 2=^H~uE{o8ץ}-[mijKִx愀H$!  \(BzhF0 ԁ.K`uA-;h/͗W?Wlthֳm?t5 @OnEN_í"qz??xW ɘGHAPz =XBpa߆vNm*. R,)WêË !p7}*hϬz,<?}^U86ԜW=a1G=j(iJq{|e9`I᳐Fx2 Pp%B:YNjwmq夊, Z Ό7s-|o303T0F* JIy5e4Z-gE?_ÿ+ >M% p#_އn:c7|Еt手DmOF7 ^۬/G?gH2z_Yovnҁ;680Q& ʰ?4'e3ʅ߾[Nv=xai[n;CpbC'*7ԶR*ď}j2eThՃeݯ?Y?-D\-J*2)FބABx n|qMߗTbP-~aۚ KjҋUrj=MUr^ ۭۧx{TA'IXeɠ 05[׀+DfM1/6=& !K,tOu! ~ǻwƈT|,7| j/nWXe"D➺L3 M;j`(4|Gmn>`S!S0MWL &!ilv7$ϖe _qLQ=|7>-eq>MG+縒_liY1+3Qfh :pDʡ2O[~!NC"y#@qG쁭DN&y=E./aoxJ | ޕڂͱ>E:o~R{)|%5)U Zb{pCgB)QBegHf~؇{,g13&=?fN2Q r\gjM"ij " +b(76l3"urD C-x%HI6%y]|ux7;tZnW$gedzUπS7O e7Y\ߌycz=Ÿ,AQ]'!ݫ{ŀݵn^p.`9V,wxx7KY[{ɱ$x -9+R 3" mOΣٝge骧Z~tէIbw{u{Jq;˔>+]uB !ft5m0nj4EgùxbG"geCqGY{w n>.f?\\PDBdFa FX8e4O 3i9[!oL}-$ȪyϿQQxBZ^*hҒFB̌D1BwT`ʁoϕGdYȇa4u08$Lk"<7IQk~WJ(jNvIqV׭*N񪍘HZ _L,Sj 0Bh#CyE[p\;a= "V` d#;1Q;uNŤ6)`H5/~S 貒]bЩ4 K8X-"/X'F) ϡG ^,q - 8!N2+j8 4zVJ$)]jRJ4ԩ̓6; ~}h#5J<ֻEa#)HP#=H0Ln jEhf x ʔy '*(ӫ0AKLTr3m3!|stQ3d7B HZxfXV-6_/+o26!%D[eB%#6zmx +yԿ.o57)ؽ~O˛Eꪸ 剀6g/UnPpcˌڛ9:El]4u6c`\#B4d n5p7ȦXx)`#EFDo$$:_ # NdS9#?x"ѯ@&)2%OSDg]蕀0\0ea2mO?a`q3|pp [~1 2Gu3Qٔ}ݑՇ|l$֚<+cx0gܕ>G$o'^IEDz`Rɞ'=: J]7)f6eד)JOڗu̇W $G ʬ ʢSoiS4NĨf/}Ѯ1B~(F7,ITk¶(Dj&O'y;()G%wd(pPkR!u"C5VʂAf.dHn^1ʢsbԚmĈ!эlJ([fN ԫ*:ʱ#EY{֠%fs)WXU r|MH91ќ)<8&Gž1" Ud5"Yis`#V&gG()9!8WZ/x]3߿LoV?n<9tqY{TT?NN55}XtIܛ;xЛ7kM %r^"?Ϋ'vs-+iE(xtjͰw6/8dstzѼ5Grt pTJpm^j `G%$s5OZ1%s_Oz+A Ή3h;ys +qy9=ӴEWHu9F^teЖ>#T^9_t/;J,_B omnښ+A( Da+ˏ|d(EZ)+0/܋" h\=GbA"&Is;bVEH4b e҂T;cQAgaI+<ȉ+/^:ٶPtᢏG$g$Sf %2DhG3qW0jI>hV RnSDa!WJR  ,o! @Cβ L?6BHQZ/$j抉31J8%# i< jcV%˸ +?vM}@.~ZT GR 0=#c戗Q0FD^:Q$J8 a$<XI!: @X3 U:wjF,ppQiL (ͱqh@}g6|4LO,g WgOCGZ>d-k!]_-S)C6e:E-zE.SчޮhƓywWg?%,ѹAtqdydzF~l&/*gN[nC(:~WmȌ;yKKr)mWg"ZqlOy'~ b͛ӂJ3-ykA58TK4hDF8Jb6(Jmk\Qw'7&1LNsw~=)ƕӳ1 Sr97AM)ly,/eizWn4dww}EHs훇* VZ5~1ظ5~%EX gϯFq<_,>`Wg-.o⹿~: C%#ݾ E\a>%g]=7d||O5Ō񾤤{/A(ke;Sl#%) ,/ fN͂2;0RmT[zkVi?x(ͤEs\z[')Xp0+Qs<)'OA{:x©9J{>H)Ǫs!nq}kWv .L>;9Ԍ(H/WE H1I0S˿^8J5M ϝY^:/&c{~E[Vmg{)s{)sS1y=8O!s.qݡ%y=|:@&i/*HbsO$Qz`Ǥ8 0 @9GҢ̄• kAsd*mb fp 8 </kQ{w^TȨsE"3 OMj:[,nqp]wS,*,*8%$3PI/TM @! ~HuIrǓkgW+-2g9XSl:N\H/,-6J{D,rbc82{?@VX E:9i9v` J sJj G W IP:g`I`0i2lC0; >IC|r 7OgU]i4bb?NVa$Yt#PE^pqA,O3ۂRPdZ 6P .9VQ V|6tq zIi+V냽 Z Szl]}$vLWNK; Id@/$Y[|?Z܌5g[ds3@(6 W -n=!01<(枤Tw<^yݝkdb. mk}nF K_RTT]y!uk_aakIj}4F4" AJڵ6!ݿ~n,oZgGW7e^K{5awҘN;10萙#"+waRC$s5,s炚 b!xz|~Rl >gF켛t-hL{ey'eԠV+nbh^{p0MTDr/jJ1: H#A~T٨d}ZU]H I4RfU&&cΘIǕ [fyjދ\Dh.s5g-\r(ɡcEivI PX81^I<, y39*wwC+s=Z߱FDwW=b+EuxQ|_(F OPRb Pn8qgϨ~c>]]DB$d !'$Fd`!QtC6$T@";5\*H6xo\+PJwep"*`)5`ߗ?TS^IRnoՍh.i ` XܧBSzYE^Ia7H&a AK=}%]BB;Hs_Zҩ?rzҩ?ZR`WU\'ZdhPm'dAG8LX܍Pk`ܑhu0 a4d] Ty=VFKQ#(чA";Yr19 #ꎤ}~;jqL)gyF\TwܝN1rEp  ts;ޝW:T\"8u<˪|dm:ut,s =;}0 [Id5#!+;̦[:J]_W˽x.~ˀr#3Ae"yfB # .'oLCjCg_>_4\D[/wΓ+U5&J=T5\Ś n2Y(/co& FҷӮlU+4 9h%D;CdȨY3@1NAY]/qq Pj1ByɭCu%DV3yu۹\/J.~XKM6( hѓW͐@Y ZjgT^ EW G.~9::\/v5j%϶aV(3]jZM!eLj)6Y]Ru-rHءoV:l[m#nI2:f0' \ätBa2(LVrY?7o::6(q312E+ak9 4 Zw}%`Jtcc`~~;iZd".8G>$)z>*=ڥfh k[nYbq i"?] \ *DƴPMDxN`S}2D8-@K0Pk2Jdnixz?[漣ZW%"y%&3Aƀ͟_Vrnp$"+gk,Ղi[Krn 8-&Y֍I/ 0}$$7"k9 Fd_~}!Vxc@)д#} %2$3sֈ@57ۈ+Bը@^ Ԁ1*b XKs8.npyGb$%ff )Icl 7]yMw;I,_>?=w\;yJJa?}r|gDh.~} f^ ,izpZ7OaYƘ?3NPfŽobzuV=iby17' ;_^6\'F_eaz}^]P Z7Z+=撷cJҜ;)PbLpQ%yl̉dbv;fn٥yG8gD13t4յ@wYyez<꼢2c(K)Ozmgju׉`\Λ.<i&r3c!;K&ytͧylō]gfSeF9` K~cK==h˝oX~Ri-^:q|mο෎}ކwR`W㋋bWټ(j73/y*MЯ+ET9c9|H_C'7A`i?(^:bfRt):L)=8o56HI_EZp=zf)sX}h$:&,*-5T6\!JU}ao {=l^̟3i}ET1!BfcbF0R`?9&8Of[쑵Uc  * |fWNJFcNIO*$;P&bG* H,e2% `-Y[5f)CFt>nw[-##Ր1TXH^⸅z@߮9 0u˰+(Rs:GBhS4%k9] L`Cq2`X߽܏6iBA_=дkGsGEt|}b8yYiJ韵IIIIYjR/:Zm Wj,ĔqGpA`Mr,'ђ)jjVھ|gVB1E i^BV'/11]) 59ror `A; Q:DZ^B3~vٷ0R:Huґ ?|O'1hNA|ұP5PcمlGn΃XۥַC3/BV//7= oJOgOS+_#y~PjS4+M)-2gwj&{A]+1BEhUbʘfA]4#{~ԑBq#%?-LGXU=O>-y?B,-X?n0bCLog].[},,PۧTN*(2ꃞjD ^+l/; ռu!_4h jQAAznL P~7o:w tC(1B{[*sx}`$y*,kR SjrM &~;m_NaQ`ɽTP3oU6_{~jt^w~h)S%i_Cbϖ|wK|a1bqsej=ZN/qr:S{1Z'N0MΓ6j],޷`2PcEKTE%˞1P\RϤ$6L2eq-PyK'4+XXtdhܞEs\1|~כ>M>F~b)٦a)>~Qި4;5/)ՋM~m߈ʔ.;M>5iM^&?FIPGiØU}@ uFLg: y]ޯn6ͮ1O}r33(@Ԥ +3J99 +H=';34 }w_aW3I%2~цD:_-%{OV D eǜCf9pw͕MRd-Ic_@۷&vzc. * mMׄ:zO>\?{WȭᗹD6%@?dIfҸ=S nV="}HIKKI,XUr*X|yxH~ʦP2F?8 H5ll\ᯮ.~*Vm_]{O*~_ݣ/F%=#lK%wKNnt(,r[AAA6yLqG4 2 CRB9`ԺeɏpxoNo{w^}LGdpZR3 !FZ:QUR+e)<ǽSyM:WUu?+y{o*9_u*wF%=c|| uwD w@?k7ywbdF|6nb ~|̨uO։ӇChHȒNJi(+G3CkqPG2Ya+&5P+@5 hB1s$׆ּw 6@<-9ֿ5)yȅ<`TX߶.JQHdi|=Z ,@i`R͌e$X.}r^7z ֬Ɛ8g@w۽ƸDBrXb!btr`qRw7}?>wǙ;뷩U C2Ru ǛRX< PwG~A C t:cG?Z/}<űn6JECBN@XM 0NQkq Ji,WA*WXE/V݆ ww`53)h^HXɓc]!P)\Aʵ[MWqVBy)6OzEQH.J@)2*_;-m4 DF[wd|dP@l3\k@L"v#BWH@{ՇsBKL^hX`䉖'|*IlRvymx"c8|,'x~ҿcx5˝ H ߎۅzh*t:JŽjףWR(}q5v4V8b4>-s:QF͈XG^Dj nUW͑)PrieoŒ.drm&g3nI ӓ(+xL2"S|q:bpWA\y9GrEBAʵ MPs4F- = 7-6Rr /^A++VẄFNX/tt! ̝9P* P@k%J&q&gV&l ~Yͧc3_![h ~ay̳TZ;5.dSf"1([gwuJCK g .觼+(՜Q0I $}V-^$@)J<%LŪ)^8/[4:\oTRd'[ K-B s#1L"${ 3T`9OM"(܍oo}* U~E 7J@zS  A!T*xܷ<{ R T(DVSˆc, sX3$|\nWѯP0~enp񠦳E8'nG]E)y>g>~Tk]BKxD#-I>]55_&kШB4b=$ !b04^#gZ%DxC:BTL0*Y$mЁ{IL֭(|$cW bTlU @%k&F+W:9)az`W@),&İN =1`2ne]82q?M0߅k3?OW>~xXl[>\.`#pɫK-S5)EnuQ}Q\ 0{Z]Mِs1kqyӍZh jtȞQ|(>(׻5N(Rɸg-+ yގDi@!њmJZe9Z; ! ; A~b^C N@Hh |,=~Aq2B Yg_XNOs2= ?*}niqp%%'!vǠ^KŝD F,d ͘o1pဟ=pc(g;I`k|kg^G=.Bȑgut7#3lkLLIr(ٵcqD%?;~,dxuսX|\_fsَ'y{>C Kr{(S }mr󊍕ΚCl3nnixn>5N9sq{ 3yN> .dq;E]yA8xO=׎1w.\]a\_3ӉGKi\q"v֮q΢TyA8̈́溚Xg)95\y #c} oO(Mc{Uw!ᙙ7B\n`3*ū; ^Ky~5#ɾVz 'ߋ mp aDV)Wv(~n(pU}8hE5,YWJ+hu2 `} 7t?ֻѶ].nTm6s u^%m"sVjx?7a \ z' {F֋{%sZ'oGpzʯo\mwkBl.  enDCp^( *I*:q[7 y=ORiґt|t(l'QIԺ@L[±\ X^DX3v;+ Q︂)GCK<8%p2ݮ*x^E9I`;?KFlE79kʇ)LEgm[Dhɝ%OSXl?K6$nCFR9DBM&~j%;K&YPQ y؁P3}=Ua+9"`ĔrQT;KN>bΒc,<(Q˃rl6'gӧڝ%/m*9"nyOs,~/);KMVzO29ϒ @qQl즍;K&$o8%Nc_3dGZe_Juc2{V[dӔ0 fj|k7il;/G.pdO h!iN4/ 2Sffh+gV0[D̊ ȴ83+7\2+,4d2/-YbҤYḙӨaEC(ps MڹC!O}'>SO8 x2{}P-dOwd,ӥe/_HOB;Êv5Sde!;.볁h.6g8$Iv~|t>M8fI; r RʵEMcqqkK_w_l8))/З,,61I13_ 9@k@>(ϹڙhY@Wu78'Ca $ a'x Sa:U !"9~xq$/= |!菻3؈'A6,xy rBY~ !L(`"3+ zˆ!8a%vH$`M_OMPˋB"bVbGw?|EhYg~KǛ7UH4&B $@Kn$eZ-J9:VJ ݟ[/{kF4r:ut! ̝9P*<  ="@`$R `+ ,KXF*O4bU Hc~bX*p`Sr.ZybL? '7? Zbx/7|ypgn!博+qz+ ?ۦN/-pKhF!ςd4 $g_u8ֶ嗑Q'1 8,k67R7 fntg}yWfU9D $gCQZX0.zYU6[:"qJnޅ57-VhgZy9UArUTD*›Ȥ4*P loI{휋F`m%8b}ЁT1 CTGt;Br'ûȂVn_l$OVYrSP9Y\$ց>ϸ\kw3ƀ{0%Yvsl>Xտѣ0<>c!d}d D{qB&fSRh@3U}vOb:O\8a)5L+ 2ƨRf^EW~w0VtGN5zՆh E #RV0A[+}Hȁ,ydX+j)ٮX[ǩրag$nj) Xɒ# 3rK:դ&d#&fFJR cȏttF`9|aDA1B2!hL2}XfPnT[AXTT3^# QHSLX"Q=E )OaT93Db(@QE@$rN֒gP8rP}A,_F%( "N7d> S\Oh LIior҇$>ٗ‡,i|C|ήo=)*}ǂ4W~WLg$VˢI/) /Jre~ $bcM1Del=?ҳ4CwfVN׎p͒)#ob([.uD'My!W򊧶n5֭ y"#Sma7nf-:&m<ݔDQںjZ6$䅋hLI$cvuӕcfhr1#:hb IL[{n֭ y"#SXobGar1#:hbMIFnkڐ.92%8jنu|A}Gv (To-{u[E4GsZUߴn{!k*HeQpsW3hi E4KvL}a>떋A}Gv*Hl2ֵnmH }ަuv5L j2ѨfzK!k]3mAB^Ȕy7v]g )GI5[w nR:ʺV5A)reT+e}u KuGT=:U^D\٭fJqjx] Js `0U x)n ] r=XYӺr hWY*kmhu'V8št֢&pDqpuV5Aq*k,ZWYkSq,*k]eUM<ZЮUI󳫬qB؋U^Be*鳫q uV5Ah*khUֺZ|v5i*k]e=M`L OܹX=-uB$qz}Lm1^q ?Oj=Yzq\Uc} F/? E DcPṮO넰iɄ*LoB]9Bhht +HO *Rj@k MQW;1B?! Em ܕSg&;AFZizt. \;B#fVᐵ,e0*u1$:kd Df0V΄ TR@eʃSͱ y~$h; E BWQ,eSBQZCB$@" ZzHEDq! !dl iȉa\b/X)#rʖaփ".2BP3a4Seޜv4G=n5䑳W lu#2^\YOq2=(Ʀ큖Y'f6vQBEC?~1 S99?{xk.]iyܞ"HaJ^ LgK};|Wq~SLgwx ׻h'іI ]ܜ+4 i̱d⣙7q#)RuQ5Gjȩ‡kPYJgiH .LsGCܓ{DŽ<5*qPyvQҀu-0,!ii g1ux{ O}{?2Ы ߕSX'fz`*M"8Ƃi1 s^7vZy^O'[ɿ)ɫJJ^=6YRKO/p4 %[NnpR^'7y} vh;4%"4з)]||v; 3lvC~ c: zv&:\3w[Q΀?Ӵ]) (};Ns@2~/FImߖûoϱA(@,8HU}%c:HyLr0[$].8kVyj ֍,(}LIbD>: JX)X(lsl1h#tjQѠ<5@85;3.@]$g2M.BXvVgǗu[ybȓ)?!xrָm1YjWZNFzJ1+5d; *ACj dc6w::mEp!1a4@/ 2>0LVhdM?d*Gt~]?)^ZrӊgߝtB0O@`n%uIWf嫾Kթx@S{;,).'@Ln3up/>K}X#e`TEJM=$n$\ɧ~ox"bSoGfCZ) f9sQ&5Tzxo1η0O)u!> ,aø(t15#oy™Q8u "+Mp`>Q Uf&pw`g& =7!b7O+G70e]1R^s`? Կ Wk07hs'uD;*1PKK={U,ȫm mCdq8`7O}?ԃcVf$ Fm1̪ˋKF@cq #XAuS0Q1 pK HCDx珷-KdkznjP| ڻoGFgx7CÝbT"nad|QGkQ^߷w.Tʺǭn4F> :O9y7q𢳐Iyk k@(|@r7 DrOi膮<؇7/ܸp-Ӳw?׀)ʂ@xZ/_lp_-?s yli{,`RH),FkY0q(N)D3X҈V`M۳ ;VExwfxǭvL`WA!p0kBhUjŘ`\͝p37#S\6t"2Ƣ3i"բ\&vy2עKJc8.Z9dwzD?OB9aչ&ƘW>1eYI}Gc|Pl'd1nCB^fɔ:yú B\ jm&Ih-{mAB^.SKTi~de(+B2{}UA{^vy=r"-n}=JRDĞD9*5 Zj[YRG׺SLHƝ, B`"gidHnA@z1(5&x@ ŕ @q! Q 1DŽA1RXahtfb-ua#5&!%0 ƫ_ c[mguL(!IzTa"vwin겼˥_RrY-\1IU7cxMv77UZܽwn'Mp4L풩~fW+s8{vi >i^ZO.7~M(P&/&?m{ҺR{taB+(迦ӵ NIܓ2A`4HL9GHaweqIz 0EyГ],vk;/;hU"Gn-/7GA2.IeE:3+###+eDh=Wg^*WSIQZ9E06Ƃ`Q}17wE/74WJc) (S L9Im-4=.US}`21] c bЕ54Jo 5Pfqkzu@~ q@/%xJH-Cfj"9j@E'B I*_m%Znmn+9g7h*$HSΏɲ~a=M$gandw "Y02fq(llęʩ4nJґ3ejFM$xίfO3"_qJ||J9#H$Jp!z'l010] ȜDrf (+鲒Vߙ4HT6H=-tZ΀ZGX[^HjV,¾⋋WlnwnusBxZ|=M#+_VF?bu&s{q7'۟Xk:#-]Nu]W:|PdIY{$"=?\ip@5hRp.(W4ƺ( !{Kq6׸Xmf!2 cl˴]< 6T| &k6_P3R?ZWny8QyDDp XPVE~h[1P+]D@sfazŖE L5 ) fOݼNA[&|E[z˜z[ZB džִd7ڐXS.Q ϝĦM@HsSc[p{˨XTJՎ|NRQi\'RE_)WNMBc]v,N%Tޮ2/{p.%x|$-V** \44k\pXTq vG)'n3~hI^W3 )WC6.@NXye5pJ*j'T1BY.{z AZI?tP0C Zd$ -8^ҳl>=8ЂE4.{hpxu9 <s>p'y,w(ayzLWyKSd*H6np)/`>

:JQk3 k#0¼cJ{)듿%j~^㻓oo׽Cv҅;t>I n S8P)Zy:msC;ґ9@.龒cvʃ(ΠVjZ4JuNA1J1N %J+X`LOK~,oNM{ KzٙA =v\]v:ˈnk!3էӯ{{8 Вsڛ֞x~4^_*oOdÒl=c&ۗ|/U.mz:P} hºCU^d0cr>O譪pqhQ1a,CA$jJHOדAThzgK8efHЯr)'QDm(64M"FH_)bJp1UnݯV MB,u]Ӿrъ`5οY`rZ׈> IF֝_>?[" TgbY"ZA,z'0o{mxٛ[>1@pU 5z=+hp noY"JYUt*TPSZ~^Rj Π귌!VlHsd@|) ٤U*#nDYJJ,Dj%l} [#@Ss0Y0pSd4K)_^^el \ 2|* 21MJ Ѽ"5_/߼sl;&}zTu#|?neaQ4iFWT2렉ǫxqsAƢoJ_[U9?C I*P_je✥u+"/USQ',s(jd7'3Ɏ7ҡAduLxf|C:0hXL)A'XׂD)'yW0cdC/BѯF\ڛ@-6W k_i*]50U^ķR rV{7Cd/ {sڱvrΈj.%Ȇugΐ !TϷs4 V00H!߼ ;ϑT7Q2rc6$}EJtnv֞1 61'^072W"G'VjnVjn[_,ZZ^fkUZCQQ4U$bQJDUPYŏv`^z2xL;N}S=fVBaHI餺)o<7ۀ1>o:W7ѣ~~~(UWE#s`05RԿZZ{:bL>{$ݻveznfVR0Cօw Wgs`xpiexd+cATxxoE k< E-OƢrT6NHwN͈u@E:3R E%T:h&sȝb@o6|P Eza 7^4s{c y7>.w eL]-w?\_]! l& цyɿYiӓ.>_m~rz^'K@2AOmWETSCvTk gHdvْӍ-)ѿ돪\j_r箠CSTT(F$yB$ mP|BZU~]]b{ B{%bG=GdNg.WQ!O(^UWl/ʆݰl7wF!PT^_VmVUzqśq\T<L\\| utjE_nI>?~P*Nc=T8P"F^E+u*y#PS't!7^: 9uYpb`AzaׂFPu(+IOhl$I mÙh@@;F\81Dfc2% }ʔR?{W6d CzgzvRX`(HXZ1 H l8HqU2:T>tg& gD}ݶov~v  ?$Od̳S|!} @0$:%+*ǡS r:6-1h9oo߾lZ4C>b:4Dh7pC ؈@# ,`.b dn.[Q).[- Bw^#rAWKƇBEړq-N2,sR30uCQ ,i@t $S@V2Q(0g#bm'1c:DX1M/7USAɶ$|&BJ[yg31 $8 coAiKRP5- p"50KH:ӀlNNG#|ġ¶K`S؆5C\>8z\)]E1ޱR봆Bz^ɊWM">IMJC'ɒka9Q2ĖA.B4eRTS!#TDŽʼon's$I_v͇Šm;׉@nYOwy:]VS&vzY Ŋj(&_W0MP'e`{ H>8#'{?K`|@f>hoDCtSם[& v:XyA8 X+Iԅ``$7^@n>AQV " KQ&syY:]miiDD:d!{-Tfn n̶;S S }!J&tOA79|sA|JP8E| Ry-iňB 7E}jٶʔm0)cR6e,2?rA/aRBIqzbEO! F!(b@Kؖڴ .ߢmٟ)?ItSa0iRj)f]Vhu!:ڭPM+!mE2 ?fod؞N 0EZcn)ccT" J ~9bde"!n˄LQT  .ai@I]|`FK%i˄$Պe?$`EE-sٶ[ʤ|ARbrwC' $ ٩vP4cBzghgak͵ɂbM ٫qy2V,;r6q:(4zld<#'otÑiC~)r:K*6"Σ}$mLCW4Yn4uظSށ:0;^LRbmPRFxIhSj1dO+..Q5v3wZJjŚ:Y<g[3"H@@ۻa)9A5Sn?E>rq9SgB:[~Oo锔 )5lG/%i{i4)g)Zjt؞ȼ !{<P]ޗ _M[f͏|U+/9`Q ; I'D2 Ӯڵ[Z;q:Oy6NWB¢:,!Pq4$ OIMd:d͵◭$Pcg\4; 㻸V*8 81 KdNtN>e;skӀYbnz$ HMb2I؝Cd&kUeyZ p^ӀE5OO4 &l>lVK? t$R {kY4Фh.Wq4E5?`J"kSFq)ۼ 6̋CuafE զCk&T~1=AUƱ<õJFUS}Գڭ=ՠX(xc\n)iw4?vvnw*DOgF5FPvYRX!RڑSa,t]X#0bXZ)Bfgɜ4`6ݮHw-_5@>n+ TSRvQJ]׵1A@1=[&zc9]oBRC :()\;#B&PLc(#l.\aзB d/`\uDxچyBS.Bo|rߧqNHxeƮXInB% <-b}A (ӆ\HL kXU)HGNV97ޏ֛ Zm9ƒpK-M- {ƕVPչi-P VB]`kyz˧V4kA$Ǹk߾"bZ;O6o H^Dbpf|SoC0q T R؃$wʜ؊fȂԇlR bܱC m$cwvљ S={6C#6Lѐĝ_fto댮a6ܙ$qԸAY9Rpb(½3t>ʿD u"xHQ3xq`ԣ⊃f' nkw< Hd8Zc{Nc| ;8IPY^RIǚ/׮|2~ L=JN><|xfz[$i>r բO*u^ ƞ{J' Yv!jN/PmoicXaVJ#V`b&D06xT|dy]9|(vydH=N\ hK#%DMHǕx],. CuYpHq\5Q;W5Fr6'/ՁH;-MEJZ+a9%3{0!1Lv9\uDkKɋjS] >\2R5Yd;S%B9څzb>cc 9!]$0(޸ -;5eCBٍiCQ#Xp<\?:_J*xB?^:~!gqzi}k}k}k.J!,D.E zZI!LkDS0gˌyWaM衯t F uNيzۄP@4Cx:& ?fIٽT)wt{2%*^bia=ֻI^~s%'AE}C&I=fU+YBG!VeتEvl%BjtR ͍Sǫ^)Ot˾<$XK^/;ϗJHUZy.G*TlNV-'5H2NO%0_KPP8Rxо `H99ILzos$yj$GYX(=+EQn.pwM3Z!֊d mj}ilU峽/ևk4'9]9 ב#-;9q `I(obr>h\~ߚ~ߚ~ߚ~?kiK Aи%,0B0`Ɛ(WTarIC@\4vsO[Դev\WmSye>@rvt)VIzf~u%r ͺXhsܤKddYE$QdRLHvHwqSB*FZ˻1AEj"G#8LqBCjg)X dcBzݱP:Y~>Ao[2@V;!dT>ZqXKvv9?C%}.To:I^|Jgs.Ǿ|_+.V젲*gGNX~*D p/_Lڎ:N<$Gx q}4lVJMo;޺g,UˆN*G ۿA\I"vgӅkX)őVcĨ:6{EPq\H|:BJ)R)e 'Ik]V<4So;vy0E ~takhD0\.d ˙e yq6)uH&O,L3%"6LYZ<eT# .fƳ_;{eMgx)T44, ѽk{ efY4@4fY=ZDtB +I3% VOύ;߹f^4bj3q n|.0J^Ex:757_>]wwxsww07`6҇RJ#3ޙu3\ߚfC}w70z UMjzΤo?64c,<™_ V~{U=b;s#R^Qxbz[ &Z<~rum_@☎|0hb\q|xˢ  mog8xĎ<<@MZ[ 5]~yfQl6Y[ے~_^e)h';h =h?7HɼyxuP} n_=x>-ˈu~F} RL]/n4axYWr~|Ig\>d?' ?>f \_zܜ'O%V~z;G']~{y;A?R 7d0UG3I|W1!Mz 6"&|tM&]4PeFtԱuQ. v<^Jɺx,)nF'I.5#/t)ɶgi4J4l='SA|ьN̿ _5"S(#-9vեܫmDϐH``XP⍟("q{%?_.CguH7NtRCUT 8>kVoB{3BkksL*S`>^̞ N ^o !9}byaKwz ߽WA`>^ֽ?/߼ tFL1m۫7hzbS7z?BPzxn<~LiTϏS>//bH'Ҧ_T|T M'(%WD#HT?N9ɭr 콯3gao֒AX2c?-F_"cU\+c F'ܝ擬?%d&1D$j:ىFt뮱sn Bu<ɔ"|1ô{*f<Ӕb:Xb }d"KIbZ{۸_Eo#HlMsΗ畨eÒӤY,p^VUufȵ6779kEqp͋`yz{ -2ӕAFԊlC~&\IEc}IϾY'azp\p`"L{}3}&>Eu<%HiI DA8HpPRh-NX%Β\}xe&K%k.]texBYDi"Bb$dLi##p ÐCa^'!6B KEp"t`1AGUzaI[-idB aSB&1(qF \"8L%8 #΅bQEC^ ?4 tOF 6XT*)R)`$a/Ò _4EV!,4ZL'Q ˔7yx3~_f=rx =Wk}~6Dޤi`sl>TO_ p~m@r.l9 ݇e]2.gygcv9rC{Ɨ?_/. ُ.k>>cj,r):rkO?ڈ/>nJ`f;[7y:d2G6mAgadw˻>`㞰0Uuӄ'r_?X`\!Y{Ċ->fpkA>?ۓ g_̝uqp0q ^n/;Vz+璩<BK;ҽ;7= S(g%0|@t_˿ ?$ xYHo#uC^FkgCVw#5S*U+wR%D4 t o@% 5pu'+/hFvUdJ]qfFÏ3cc/їa3~?Kеl tO6'rQRjRk/B` ;;I׸H4/n(gj}`_ofo>~d? ˏ8; \q/y[V :PCeH X:8ɫt_v3 ,r֡E<ٻQ/0Qhɚ}Eݸ懧?5W]{XkMH[a4u ؂.(ԼP! kxKtƍ;rOѹv:%T.B :P9՝K<u2w7+ ?/CvtB漗vsݺ#Վ|/T5>}i^}v)s&X:GY݌\R3q ƛr8o.Ϛ`[n;47oyO0qfSלĪeVy6_\ sC{nO}p~l& 7ǤX0'/RMMIM*JGDh\loz&FbiHF3MyLymίcmpsw)],/ rcŵ4||tuZ.=yMús6)$)*['lV|ZM9F.u+np [ @(^iݛU:[$٬*9`3=nV@b|k7% HJGcIW'ߺ"_IP(8dŗhk%T<nU1F/+E1)vQlh-\'׹>\-`bV֙U'x!`摘}欑G7 /LY_Ćy8_hU2߄ɪ^?x4adF«\*-d>̝(?MO 5aAh64MÇl {K67fBߩkLw| j-#o,u,08J[D3I0=x9psBVH #pq9L͏/aj'_:s"Ǘ9'W'U)"d-P1'j1Z3E'G t_~qb|<]re%+X+41l80VR ٗk1X@>C [csfi V ܶ>!1CkY`Ra}QX0\R*v {Mn~ L@Z=2- FU+ģ#`]ƪ?[<ف2ǣՅ=9MV]&4W}:7^W+Ӳm5rxDH Y$0 e8840Q0E#^&xGx*U1m jGidc e{4p, XqXUS$݌QX$iٗ&[:{=͎S@ɏjtz:EXW&de aȐ(%n'7S.q:[c \؝'|AIF+QO}Y|-d>p3ꦂSKIFSJU^it6:%A)v(nSSpيV'ekR;m y1"Eݲo(^MSt]*n;?^(WApA'Z?cӋ#T#c贓SGF8:eb5@ƄFYaiÂt,T(RmpIjG=+ORRV x? _̋''UM_gvrfh.{Ll3QF=bS0$MQn cJHB nLĂXkv}fͦ}bͦ5J4rITWGh=؉ub'N.8Dj.DD"D MtCkBS)Bb#p ̰ܨ0A^Mے VKݥxVN b(gT&\ʔc*%2ŠiCbu),@nNx}nH;N$SRƤƈP #X$M tSZun0!DPggAP0BmocG+Z4jVxZV8|O;;|M;#*.7 TNs~o{jŘb[+'rzcqYrq̄W =4ƒM,"Kvj$]tiw E`ȵ/w?8{?.%1s9~š=1ڏ.k>>cj,dfa8,4O?8'>ne.mbGGa"*7Tlߪ#%m J츚FE|'Z ц&Ά0L: P25Ld Y1͹ (թL5D2.4p |I ub Q[a]Ecc)&f+|8od{g(?2@Ym)'J !z$_89C(w㮓gz6_!rJE)A|MveKhZ~fHZC\$L߬5enkVj@GI7^cOfdXn*RCޛս[Higһ6r4..0mާZy) Q53‹_o-;O[ _DsTqNIhjޗ'gp-ӗ_7:`f 3DAk;I%["@yHI 3|}I9i _6GI<6"3N,(vbao|q!g@ZNN[6p_zgOD˯Et4~KRO gNũS3gO;6db=Iy!<1`oy~c.r,lm϶p)ܱ yFIzS͐z}`|n 8F DU t(ݱR ut6Odrb-vf{}{#u ÀZ=xZBm3ZՊqL% fpw'NQ%d |y*DށwI uO@UU}2sw;dv(stKGK+3Bd:lNYUç0X/M>^|Wh)#QTkg -&6Ik 9]wVF:jǐ! ^<>$b>Y֊J$k3M5'xEj.#~.jy84O;Hikq]x)evJ]Rhw=a4rh^ȝkReq̡ @W9YRyFVjDDXjNU2 YʔtC 'N\1[Fsrb{W:lꎘJd2SSxIOtS2zm'xX\IG}Wt_}}AՇ|_B8M: !A,!4%gŒa!D0dD /'L{ޝK6Q<L:%;dsx[n|TM 1f*JʲaEL"(BqXا!aQBV`4BQ,! a_ SA$Ũ wX"rpTTԺX-x ߻ O@*ѻGV7@~ݻI(ReO}7?Cwg=vc s맡i`&$tBbUxO}˿V!HS뚡UHz&[ByEiA2V9`i+f+WT[sArU{R3îC8ũb9?Y_r`eMsq`Z81`Sۭ#NT~ݒ8ɂ8..W$ad(&?خO3,Zgӿb~p/snpM @^9!r@bV{9 siliڥ wś9OݎT$/ya-1MP!NϦaz_}jo_&*UC}8Y#35uvj׫yg̛|`rݠ$'2_8XSU>;߯;tf~ AziA3\i|9*iû} )%ƈph7_sh -F䍦mTEjUUS*agP((JCA }.D4 HRPss (&;J!4T˳cr> қ*^L*蓽$՛x ߻ Ic)0ѻ P{w{}#w={Tӻ{ =YzޟƋx`wodzF[‚vL*u(Fc)uRAsJE}a׌ U/ٔL$zA*m秤7zoV]E=*{\$/ްVzDsPFjU[U:ɳ+}k]Pzo}-K8u'ݛ@XNB,m'ӻ LzLz4zj}9@1Ǥ5q`REgQcQ[G[- . olbE $ y \]W'f!r64 X Ta7QԹuE^.h̀@CđcS{~3! ؼƶG$syܱ uh"ǫׇ ^ވ7W0qzoˊ&<@E7.S̕T4..uwq*$"k*'5 ZB@E4X`i MsvNUۋ'*vh?0,<]zv41XU긳f\58Nq`"e<A%VM0XzbAr?];8F J^QXkUIՙ85{*8HV4tklۑUxN^ccne6u+) .svهJ5VVb&'C1Q-)`uv}Bpa:h{E_uhT\9R@œDžYir=27rB0}GjnPOhϓT*/\<7G:Zx!ҝ2H@&˙Vrȴ&m5n'UNByK{Oڕt/rTWMwCP 经f3F.!F͓´DZrMfk׳dW9d\gN&ٻ°"H\O{EߟQzДU% N|=e+[]2kJKLhBw/`n8u`b!M'n]Nuh<-7ŋ8`O|i׎zc}8]>,c}x+Xb`'ǫIT '& 2Dj.,YEXnY<ΐd,ϯ4g] ~vO# a O^|`|=gP|3Xfr#?@:'xqpyE[9鷹HkK[»8':7.xTWC%޴6R(4J e rd˓"Ĝy ;$u=ErXU%Fv< d"kcd)'~ 58 oˋqz?Ƴ1*E hA\|O?3G_p=U;3؛S}{;"'a[?lӋPi["/?O%?~_43#thVn?}p՟&$CNtv]n9Atʹ9g@(%?]PUU.Zd #Z+4wUU\b{Lsͫ; RۛꬭLVii+n` "6QՓ+L^a KcpJ!e%믢a#Q>>:~Vgд~nOXbi&l,sEC,5LPДi:~~z`%FtF[!J] %#-D.BwmdLwIHj1,d>#qpu|rRU."$CxfVZ̽@)q/lG^ؑ g[G8 *.;!1hd+@A0Xmr"G%@!&DQA8gd4 塠SEl :B.K TtU_,{^vҌ7D_:>?ͳ|9->F |*>__:M0A6:St%i!;2M2ӔvMO)7NWuRⰉ'3vI:5|D/: C!&(ǙlQC.1DR6&ZCa7i_[wotx>DՑ;c!6-A l0yuFTE10-PArP3}Iz M! JdWte%Zt#Ygunɧ<8a[q,D6,DV|{EL8R%* ]* TYiV_jv83lF?!(X"[ a-i Rzb +^J3#Vs ha6 ZCvȲeQ`Y$D*k{dYa145s!jGlQ R ٴTOcbSk {^I7uMZ _pF+BvB$ۖJ}²,,/TݱF9آ$e|),l%C*Ye % +j3MMc ZIcd t ?X횠G0odQ#æ{&fVPjTo[cHYVĈ- @lHF`ǎv~k)4yd\A|R`IPj~ BR1 [*6$k4D cIV#k /=@=@m1Rd Ɉz =⋊A 5 D$Bۄpsrt\ФT#  52dsi7uώt((ӻrs06^F]]SsWwW~s;'NS}= 2+4~d }?&߫NJ>7^ ܿխ68| ~9stϙTn[yq Q ^M&iwKn}|(:Ɔ5M!/XDd95j5ZzXݹYlYr{ctV 2i!CRJы,cdk^t1%G\ӺBZ7p{`'q0 nF:V;1pKR2RT/V!O/k$ _acN>vЎWQ(M.ؘy̲zD-w%X2F^J(@%,>u1hQ?,LzY ˲! ,vfa6)KMMȪz:W&ҁKжt{)k Кc0Κ=FcjSJc\݈ªLc ԅ55"#mZc:[nrp8BV,u9v0خ$ svp>7-W*_gVJz9x: 8:\I E{RM@y5 tn؈'gQAXY ] z[XRE󾝘":l1ui_׋݃gX;vjt;O^Rka#KI`Z}y6-*v$!C/~;K,D' d7]?߻1.K 6ŐE]%lFEYE;9kڴԶ^@cQ[c$\H'4BJdB"S/͙mb/ʹD 5> .gڧt72Hߏ޻S.ի_| +ٱ?h_̾= o}~\wo4u7^ Lq2QV'Yhx1DrҼuߗ%ʌ;BiٱLzmw*46:փZE4^&YՏ)LOw\vWl&ͻA!y%.h ʘd EH("BtQRʁ4B ؀**hr kI#y9x0zk{iJlXvu3MNY/a70h]S(8b y,ze+c1 ۊ}YSI_n֗6!;Oiv]nwuh+'liۧ2NEuxj[/\?'T-:߻wcMLwzݫ9Zi l*@5$l%5H[/9)|NNtR! 쭪}l41&NX2K*~ WЗYώ $+}:Sϝ{8_DiMOFv pbJk7]@tE4}xO2Jjm Lihet@[l p$%BV3C}y0;b. #wP FY,_9.s+vT \=W~r>?;8cSJI/6v+D *9gU[G*!ya[?}~=W"{4HOMbuZ7yp)mj~g-P`,l)Ao%w H{QK?l~e+985hւ^h,;Xi>dY+nMd ;3O~jra2f;}wiΊc$w,ª9`2$3瓋U,nQby":﨣Sy/j6$䉋hL8xDa[,BD'u[TwKj6$䉋L0cbLz~, .dsYpMPI1~/er{kD)3`t&DcR qe2d CKHIPʄ9 I6H kfjҢOb!0%^a]phW_5Aqi |ǥ,l>.Q8'gޱ  {4˹c/f JQ]Vs N*ޥmwԝCJ2J G>XupMαL@*yhtnMOfM- Wts@Lخ *,* عϵŒ0Acna=Tkwd8՚aMyjB9G<!w~JwhG y"#S S&˃щ}G>KNnmH#ҁ2] τAl㻗/k%‰y{7C˓ІTjX2r{s%8^!j6p^OB_E쟮ƻz~wnI؛[{[t،gX>z={w[4mxa_pE>A?SHWHi']5)T@CZy= zH39zhgˬ"9_MWaEOE{[D>66!s($(}-P)O>  (: %żc0Wt?ShC/z=Y~Pv8 ^YC9ֵDΐo^~Wvcz3l,ݦLeH,3::>$Ms3I]AYc1xW`;(-;eS8]V'I?pjQcR&TG%!ҕP~:,EYYLJ[*PIfB$))LDY,#XkYO%n_,EFVC/_itZhC$䉋hLZ'MSlFɃNߑՌS`Ꝉ/ތ>6Oh3$䉋L@b)f ?a.Py\Q6yʧn.KbxSd]mʮ IPv|~bnDԿT?5Kڥ1'.߭K,3ق;2e`ElW/^`"+8 eZMlBhl \|v" :ɜ)/OeQ7M[{T+Y4I)X:Є%6(tLI^;b|Sj9@f4> a,Y |]&K,- < ĢT>J? |d)zaTѥ=!71%ŵbDR)0m>~yC=;< = qAM-2\!d8Ì" 9V(OJ$ e8Uy2c [8 !~pU݉l~s|NWb0^@Dh8!:Bb!B3uM)5x>Teí|n z1A\vL(G=͙QY!s&RVLik8" PҌQJa$~MS1nj@… 11640ms (00:06:07.502) Feb 21 00:06:07 crc kubenswrapper[4730]: Trace[886934269]: [11.640634822s] [11.640634822s] END Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.503154 4730 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:07 crc kubenswrapper[4730]: E0221 00:06:07.504838 4730 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.522650 4730 trace.go:236] Trace[924890170]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Feb-2026 00:05:56.995) (total time: 10526ms): Feb 21 00:06:07 crc kubenswrapper[4730]: Trace[924890170]: ---"Objects listed" error: 10526ms (00:06:07.522) Feb 21 00:06:07 crc kubenswrapper[4730]: Trace[924890170]: [10.526612583s] [10.526612583s] END Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.523070 4730 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.524593 4730 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.525852 4730 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.526525 4730 trace.go:236] Trace[1869033455]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (21-Feb-2026 00:05:56.100) (total time: 11425ms): Feb 21 00:06:07 crc kubenswrapper[4730]: Trace[1869033455]: ---"Objects listed" error: 11425ms (00:06:07.526) Feb 21 00:06:07 crc kubenswrapper[4730]: Trace[1869033455]: [11.425926161s] [11.425926161s] END Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.526558 4730 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.530672 4730 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.540541 4730 csr.go:261] certificate signing request csr-tg7q2 is approved, waiting to be issued Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.577318 4730 csr.go:257] certificate signing request csr-tg7q2 is issued Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.596653 4730 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": EOF" start-of-body= Feb 21 00:06:07 crc kubenswrapper[4730]: I0221 00:06:07.596723 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": EOF" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.225564 4730 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.226001 4730 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.RuntimeClass ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.226011 4730 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Node ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.226021 4730 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.Service ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.226073 4730 reflector.go:484] k8s.io/client-go/informers/factory.go:160: watch of *v1.CSIDriver ended with: very short watch: k8s.io/client-go/informers/factory.go:160: Unexpected watch close - watch lasted less than a second and no items received Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.362367 4730 apiserver.go:52] "Watching apiserver" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.375135 4730 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.375687 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-wkkq5","openshift-machine-config-operator/machine-config-daemon-r8q7m","openshift-multus/multus-m9nnd","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-ovn-kubernetes/ovnkube-node-rjptr","openshift-multus/multus-additional-cni-plugins-srs5q","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h"] Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.376233 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.376323 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.376345 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.376563 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.376789 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.376865 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.377101 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.377630 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.378138 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.378025 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.377661 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.378218 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.378325 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.378591 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.381971 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.384282 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.385623 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.385945 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.386204 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.387182 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 17:25:22.44879465 +0000 UTC Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.387345 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.387449 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.391100 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.391671 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.391998 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392155 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392208 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392547 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392762 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392835 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392939 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392980 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393033 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393084 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393181 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393225 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393303 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393333 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393406 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392768 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393471 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.393187 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.392988 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.394245 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.394475 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.395262 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.420302 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432228 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432285 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432317 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432346 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432419 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432451 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432476 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432498 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432521 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432576 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432632 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.432655 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.433301 4730 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.433611 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.436334 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.436628 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.440997 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.448261 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.450626 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.456152 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.456201 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.456219 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.456292 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:08.956268924 +0000 UTC m=+21.247104654 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.457609 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.457636 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.457648 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.457691 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:08.957679443 +0000 UTC m=+21.248515173 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.466652 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.474202 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.478133 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.479307 4730 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.485537 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.519458 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.534298 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.534365 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.534408 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.534435 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.534462 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538017 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538370 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538413 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538459 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538505 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538532 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538558 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538584 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538619 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538647 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538680 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538706 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538735 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538780 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538811 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538835 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538864 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538920 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538946 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.538966 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539049 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539084 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539110 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539138 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539308 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539346 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.539378 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550030 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550302 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550358 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550401 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550445 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550486 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550523 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550556 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550594 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550624 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550648 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550672 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550690 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550711 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550737 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550756 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550805 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550841 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550878 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550929 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550962 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.550992 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551014 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551040 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551071 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551097 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551126 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551155 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551186 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551214 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551242 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551269 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551294 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551342 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551385 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551413 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551442 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551474 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551498 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551526 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551555 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551589 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551616 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551645 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551674 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551700 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551730 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551755 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551781 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551807 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551836 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551862 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551890 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551944 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551975 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.551999 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552026 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552054 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552081 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552105 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552133 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552164 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552186 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552213 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552240 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552283 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552316 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552346 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552376 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552402 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552431 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552461 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552492 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552522 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552550 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552575 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552599 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552619 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552639 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552656 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552679 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552704 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552722 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552743 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552767 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552786 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552805 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552827 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552853 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552878 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552927 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552961 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.552985 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553007 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553029 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553049 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553070 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553089 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553111 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553130 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553151 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553172 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553202 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553225 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553245 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553264 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553285 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553308 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553327 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553346 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553367 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553390 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553412 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553434 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553456 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553475 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553498 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553518 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553537 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553558 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553582 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553563 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553603 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553627 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553555 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553652 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553679 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553704 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553725 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553749 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553770 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553792 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553816 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553839 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553868 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553892 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.553919 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554056 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554116 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554161 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554193 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554233 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554269 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554304 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554336 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554337 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554348 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554377 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554509 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554590 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.554630 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:09.054533499 +0000 UTC m=+21.345369229 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554710 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554719 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554805 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554816 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554857 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554939 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.554969 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555000 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555039 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555064 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555135 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555163 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555188 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555213 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555236 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555260 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555280 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555302 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555323 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555357 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555379 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555465 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555533 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555544 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555585 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555599 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-system-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555632 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-socket-dir-parent\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555655 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-k8s-cni-cncf-io\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555747 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-system-cni-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555778 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555806 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555855 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555861 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cni-binary-copy\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555926 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-netns\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555952 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33513de3-5480-4aef-87ff-879f9e7a475f-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555974 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.555986 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556012 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556046 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cnibin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556095 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556144 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33513de3-5480-4aef-87ff-879f9e7a475f-rootfs\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556181 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jdnh\" (UniqueName: \"kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556206 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-multus\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556224 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-multus-certs\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556273 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556294 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33513de3-5480-4aef-87ff-879f9e7a475f-proxy-tls\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556313 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556334 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556325 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556375 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-bin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556393 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-kubelet\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556419 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-conf-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556465 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lmvms\" (UniqueName: \"kubernetes.io/projected/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-kube-api-access-lmvms\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556485 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-os-release\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556502 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556545 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556569 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556589 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556614 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556636 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-binary-copy\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556658 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556676 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cl57\" (UniqueName: \"kubernetes.io/projected/33513de3-5480-4aef-87ff-879f9e7a475f-kube-api-access-9cl57\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556715 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-daemon-config\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556753 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-cnibin\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556785 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556821 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44z99\" (UniqueName: \"kubernetes.io/projected/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-kube-api-access-44z99\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556843 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556866 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556938 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z762\" (UniqueName: \"kubernetes.io/projected/638e980f-5038-4a5d-85d2-4aae577964af-kube-api-access-6z762\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556967 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556993 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557018 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557070 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557100 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557126 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-hosts-file\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557178 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-os-release\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557206 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-etc-kubernetes\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557230 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557255 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-hostroot\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557282 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557308 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557331 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557455 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557479 4730 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557495 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557511 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557528 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557547 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557563 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557578 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557596 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557611 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557626 4730 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.557641 4730 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580370 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580587 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556580 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556755 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.556938 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.558299 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.558797 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.559033 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.559231 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.559839 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.560470 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.561041 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.561381 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.561408 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.562459 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.563242 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.564240 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.565588 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.565653 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.566805 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567094 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567170 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567345 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567341 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567697 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567808 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.567985 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.568762 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.568772 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.568844 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569051 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569102 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569272 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569296 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569380 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.569491 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.570546 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.571387 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.571581 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.571862 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.573794 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.574162 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.575303 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.575591 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.575989 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.576020 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.576278 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.576647 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.577127 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.577361 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.577542 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.577628 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.577840 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578012 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578201 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578213 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578405 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578453 4730 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-21 00:01:07 +0000 UTC, rotation deadline is 2026-11-14 03:33:15.559160403 +0000 UTC Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.582534 4730 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6387h27m6.976633179s for next certificate rotation Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.578756 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580097 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580506 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580523 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.580539 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.582642 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.580896 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.581086 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.581609 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.582079 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.582113 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.582836 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:09.082809932 +0000 UTC m=+21.373645662 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.582890 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.582937 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:09.082924915 +0000 UTC m=+21.373760645 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.583444 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.583609 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.583892 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.584006 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.584278 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.584897 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.585451 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.585539 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.585866 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586109 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586128 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586338 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586510 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586582 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586821 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.586856 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.587096 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.587342 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.587399 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.587407 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588180 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588209 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588382 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588660 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588744 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.588781 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.589537 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.589777 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.589800 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.590111 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.590612 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.591376 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.591607 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.592146 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.592372 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.592688 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.592767 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.592960 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.593059 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.593180 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.593247 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.593313 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.593328 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.591896 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.591201 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.594671 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595232 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595327 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595486 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595523 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595668 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595776 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595830 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.596097 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.595978 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.596372 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.596424 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.596576 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.596620 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.597424 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.597894 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598116 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598260 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598287 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598259 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598166 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598300 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598396 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598651 4730 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598725 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598800 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.599088 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.599279 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.600761 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.600778 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.600645 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.601368 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.600556 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.602007 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598535 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.598591 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.610113 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.609964 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.612916 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.613151 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.613308 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.614447 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.614541 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.614575 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.614956 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.615208 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.615425 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.616201 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.616513 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.617200 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.617473 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.617724 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.619260 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.619994 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.620207 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.621121 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.621161 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.623214 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.623268 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.624277 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.625200 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.625279 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.625609 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.626232 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc"} Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.626262 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc" exitCode=255 Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.626592 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.626890 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.627061 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.628893 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.628915 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.629168 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.629267 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.634978 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.635411 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.636688 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.655520 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.655530 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661482 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661531 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661547 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-hostroot\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661566 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661582 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661598 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-system-cni-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661636 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661652 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-system-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661668 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-socket-dir-parent\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661689 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-k8s-cni-cncf-io\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661707 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661724 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661740 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661757 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cni-binary-copy\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661773 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-netns\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661798 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33513de3-5480-4aef-87ff-879f9e7a475f-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661814 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cnibin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661831 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33513de3-5480-4aef-87ff-879f9e7a475f-rootfs\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661850 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jdnh\" (UniqueName: \"kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661867 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-multus\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661884 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-multus-certs\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661919 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661940 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-conf-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661963 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33513de3-5480-4aef-87ff-879f9e7a475f-proxy-tls\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.661989 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662007 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662027 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-bin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662042 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-kubelet\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662059 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lmvms\" (UniqueName: \"kubernetes.io/projected/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-kube-api-access-lmvms\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662077 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-os-release\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662094 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662110 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662133 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662149 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662138 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662195 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-multus-certs\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662168 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-daemon-config\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662260 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-binary-copy\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662282 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662301 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cl57\" (UniqueName: \"kubernetes.io/projected/33513de3-5480-4aef-87ff-879f9e7a475f-kube-api-access-9cl57\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662342 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-cnibin\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662358 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662373 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44z99\" (UniqueName: \"kubernetes.io/projected/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-kube-api-access-44z99\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662400 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662415 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662441 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z762\" (UniqueName: \"kubernetes.io/projected/638e980f-5038-4a5d-85d2-4aae577964af-kube-api-access-6z762\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662461 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-hosts-file\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662479 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662496 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662517 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662552 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-os-release\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662568 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-etc-kubernetes\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662676 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662688 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662699 4730 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662734 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662745 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662754 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662764 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662774 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662783 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662793 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662802 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662812 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662821 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662830 4730 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662839 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662848 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662858 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662871 4730 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662883 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662895 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662924 4730 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662934 4730 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662964 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662974 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662983 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662992 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663001 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663013 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663012 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-daemon-config\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663053 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.662848 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663022 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663098 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663103 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-conf-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663110 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663121 4730 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663132 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663141 4730 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663152 4730 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663162 4730 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663172 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663180 4730 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663189 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663198 4730 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663210 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663220 4730 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663231 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663242 4730 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663254 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663268 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663278 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663288 4730 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663300 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663309 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663319 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663329 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663338 4730 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663348 4730 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663357 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663365 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663374 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663382 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663391 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663400 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663409 4730 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663417 4730 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663426 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663434 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663442 4730 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663450 4730 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663466 4730 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663477 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663486 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663495 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663495 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cni-binary-copy\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663503 4730 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663083 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663521 4730 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663534 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663545 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663549 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-hostroot\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663562 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663568 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663573 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663589 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663599 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663607 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663618 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663626 4730 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663636 4730 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663644 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663654 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663662 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663672 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663679 4730 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663688 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663696 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663707 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663715 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663724 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663734 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663744 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663753 4730 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663762 4730 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663772 4730 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663780 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663790 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663798 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663807 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663817 4730 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663826 4730 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663836 4730 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663844 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663853 4730 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663862 4730 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663871 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663880 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663889 4730 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663898 4730 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663920 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663929 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663938 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663947 4730 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663955 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663964 4730 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663973 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663982 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663991 4730 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664000 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664009 4730 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664018 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664027 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664036 4730 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664045 4730 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664054 4730 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664063 4730 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664068 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/33513de3-5480-4aef-87ff-879f9e7a475f-mcd-auth-proxy-config\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664072 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664089 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-system-cni-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664098 4730 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664108 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664115 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664117 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664132 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664145 4730 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664146 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-cnibin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664153 4730 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664162 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664172 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664180 4730 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664196 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664205 4730 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664214 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664223 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664233 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664241 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664250 4730 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664259 4730 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664267 4730 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664276 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664285 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664293 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664302 4730 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664311 4730 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664319 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664328 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664337 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664347 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664357 4730 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664366 4730 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664375 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664384 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664393 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664401 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664410 4730 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664418 4730 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664428 4730 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664437 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664447 4730 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664455 4730 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664464 4730 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664473 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664482 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664491 4730 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664499 4730 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664508 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664516 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664525 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664534 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664544 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664554 4730 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664563 4730 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664572 4730 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664580 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664171 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/33513de3-5480-4aef-87ff-879f9e7a475f-rootfs\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664626 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-system-cni-dir\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664656 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-multus-socket-dir-parent\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664658 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-multus\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664683 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-k8s-cni-cncf-io\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664686 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664708 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.663597 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-run-netns\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.664738 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.665312 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-binary-copy\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.665577 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-hosts-file\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.665845 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-tuning-conf-dir\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666013 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-cnibin\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666350 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/638e980f-5038-4a5d-85d2-4aae577964af-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666440 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666540 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666593 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/33513de3-5480-4aef-87ff-879f9e7a475f-proxy-tls\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666628 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-os-release\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666642 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666667 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-etc-kubernetes\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666675 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666746 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/638e980f-5038-4a5d-85d2-4aae577964af-os-release\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666789 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-cni-bin\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666823 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-host-var-lib-kubelet\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.666977 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.667091 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.667121 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.667201 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.667241 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.674377 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.674751 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.677729 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.685635 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44z99\" (UniqueName: \"kubernetes.io/projected/ac51a893-4a0c-44ed-8284-0aac9e8d02ef-kube-api-access-44z99\") pod \"multus-m9nnd\" (UID: \"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\") " pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.686280 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lmvms\" (UniqueName: \"kubernetes.io/projected/d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14-kube-api-access-lmvms\") pod \"node-resolver-wkkq5\" (UID: \"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\") " pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.692242 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.692601 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cl57\" (UniqueName: \"kubernetes.io/projected/33513de3-5480-4aef-87ff-879f9e7a475f-kube-api-access-9cl57\") pod \"machine-config-daemon-r8q7m\" (UID: \"33513de3-5480-4aef-87ff-879f9e7a475f\") " pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.693152 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jdnh\" (UniqueName: \"kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh\") pod \"ovnkube-node-rjptr\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.694236 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.699089 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z762\" (UniqueName: \"kubernetes.io/projected/638e980f-5038-4a5d-85d2-4aae577964af-kube-api-access-6z762\") pod \"multus-additional-cni-plugins-srs5q\" (UID: \"638e980f-5038-4a5d-85d2-4aae577964af\") " pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.702075 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.702675 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.711078 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.717244 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.722661 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-wkkq5" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.735266 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.735756 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m9nnd" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.743494 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.747780 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.765628 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.768211 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.768895 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.772536 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac51a893_4a0c_44ed_8284_0aac9e8d02ef.slice/crio-e8296d1bee0b2bec6a7f08a9c679ef97df3c193a498eafb2213b73b0429e275a WatchSource:0}: Error finding container e8296d1bee0b2bec6a7f08a9c679ef97df3c193a498eafb2213b73b0429e275a: Status 404 returned error can't find the container with id e8296d1bee0b2bec6a7f08a9c679ef97df3c193a498eafb2213b73b0429e275a Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.782936 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.791726 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-srs5q" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.795289 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.813164 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.826928 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.839323 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: W0221 00:06:08.848075 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638e980f_5038_4a5d_85d2_4aae577964af.slice/crio-abc29c84e1246ebcb944625fbe40cf494d531591f6da4058fca0196ee99ae08a WatchSource:0}: Error finding container abc29c84e1246ebcb944625fbe40cf494d531591f6da4058fca0196ee99ae08a: Status 404 returned error can't find the container with id abc29c84e1246ebcb944625fbe40cf494d531591f6da4058fca0196ee99ae08a Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.848265 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.862765 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.876503 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.888953 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.911176 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.932454 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.945582 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.956962 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.966929 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:08 crc kubenswrapper[4730]: I0221 00:06:08.966973 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967119 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967140 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967154 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967176 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967197 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967209 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967213 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:09.967196171 +0000 UTC m=+22.258031901 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:08 crc kubenswrapper[4730]: E0221 00:06:08.967260 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:09.967228012 +0000 UTC m=+22.258063742 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.067698 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.068119 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:10.068094541 +0000 UTC m=+22.358930281 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.169377 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.169444 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.169555 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.169618 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:10.169605187 +0000 UTC m=+22.460440917 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.169658 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.169676 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:10.169671379 +0000 UTC m=+22.460507109 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.388359 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 21:11:11.122627487 +0000 UTC Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.630514 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"7dbb8ebae9d29aa2e1e138bc5773a70043fef95288f7a2d415fd6012cff96d44"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.632583 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" exitCode=0 Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.632658 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.632680 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"90db4f07cf055e32c79857260824754a4e4f6c0f7ae993e96b137c91cfd994e6"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.635473 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.635509 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"0a549021efaec2704c42ba7e4eb90b69523fb20a0ea67ab0ff47c9b1fe0f3405"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.637990 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8" exitCode=0 Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.638126 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.638174 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerStarted","Data":"abc29c84e1246ebcb944625fbe40cf494d531591f6da4058fca0196ee99ae08a"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.645833 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerStarted","Data":"71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.646135 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerStarted","Data":"e8296d1bee0b2bec6a7f08a9c679ef97df3c193a498eafb2213b73b0429e275a"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.648782 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wkkq5" event={"ID":"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14","Type":"ContainerStarted","Data":"edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.648956 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-wkkq5" event={"ID":"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14","Type":"ContainerStarted","Data":"61306da2d1d9adf04cee9440a7118261fb292f21489d8aa55635be3ef072a687"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.652691 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.652753 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.652766 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2b11a18b934679e047a61e3aa9cb0e0921bf4037b1fa261ecb6fc1ae46b86ae3"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.656631 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.656721 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.656743 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"6efe8c5a6e6a7cf2d96d9f811af9642a7693523110a512eead80806bb0ccab93"} Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.657575 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.663977 4730 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-apiserver-crc\" already exists" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.664277 4730 scope.go:117] "RemoveContainer" containerID="e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.688695 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.697257 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-f7kbj"] Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.697928 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.700729 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.702614 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.703488 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.703518 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.704964 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.722722 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.741084 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.769558 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.777151 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a502eb4e-6f7f-4567-a7fd-bc15eb806845-serviceca\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.777202 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a502eb4e-6f7f-4567-a7fd-bc15eb806845-host\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.777253 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-78bbl\" (UniqueName: \"kubernetes.io/projected/a502eb4e-6f7f-4567-a7fd-bc15eb806845-kube-api-access-78bbl\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.794310 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.818851 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.834543 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.850035 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.872720 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.877743 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a502eb4e-6f7f-4567-a7fd-bc15eb806845-serviceca\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.877784 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a502eb4e-6f7f-4567-a7fd-bc15eb806845-host\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.878006 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-78bbl\" (UniqueName: \"kubernetes.io/projected/a502eb4e-6f7f-4567-a7fd-bc15eb806845-kube-api-access-78bbl\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.879134 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/a502eb4e-6f7f-4567-a7fd-bc15eb806845-serviceca\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.879192 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a502eb4e-6f7f-4567-a7fd-bc15eb806845-host\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.896221 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.909804 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-78bbl\" (UniqueName: \"kubernetes.io/projected/a502eb4e-6f7f-4567-a7fd-bc15eb806845-kube-api-access-78bbl\") pod \"node-ca-f7kbj\" (UID: \"a502eb4e-6f7f-4567-a7fd-bc15eb806845\") " pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.923546 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.942957 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.968166 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.979757 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.979847 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.979994 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980039 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980055 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980136 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:11.980112454 +0000 UTC m=+24.270948194 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980181 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980235 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980255 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:09 crc kubenswrapper[4730]: E0221 00:06:09.980342 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:11.98031455 +0000 UTC m=+24.271150280 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:09 crc kubenswrapper[4730]: I0221 00:06:09.993766 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:09Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.029770 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.048851 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.051002 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-f7kbj" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.073497 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.081264 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.081433 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:12.081400985 +0000 UTC m=+24.372236725 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.119480 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.133535 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.149146 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.166389 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.182307 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.182389 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.182486 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.182550 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:12.18253219 +0000 UTC m=+24.473367920 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.182549 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.182667 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:12.182639804 +0000 UTC m=+24.473475714 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.192082 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.210375 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.389021 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 15:26:31.493293997 +0000 UTC Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.433964 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.434054 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.434103 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.434069 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.434289 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:10 crc kubenswrapper[4730]: E0221 00:06:10.434442 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.438339 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.439713 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.442496 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.443981 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.446473 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.447783 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.449303 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.451577 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.453023 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.453954 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.456506 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.457725 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.459853 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.461011 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.462211 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.466344 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.467694 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.469308 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.470347 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.471334 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.472716 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.473620 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.474963 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.476035 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.476764 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.478430 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.479494 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.480796 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.481810 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.483189 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.483972 4730 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.484199 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.487081 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.487675 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.488194 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.489758 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.490867 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.491488 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.492546 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.493308 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.494281 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.494980 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.496014 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.497089 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.497610 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.498533 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.499151 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.500301 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.500852 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.501402 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.502269 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.502847 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.503853 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.504549 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.664410 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.666927 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.667354 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.673089 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883" exitCode=0 Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.673162 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.676923 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.676973 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.676991 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.678357 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-f7kbj" event={"ID":"a502eb4e-6f7f-4567-a7fd-bc15eb806845","Type":"ContainerStarted","Data":"57f8592ac9ed88d53949f19f15c2d2416f2f1734e42e2283781498fa5db9b0c5"} Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.685949 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.702191 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.723793 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.756287 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.779018 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.807059 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.823216 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.836052 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.855965 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.885516 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.899250 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.922880 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.945375 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.968822 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:10 crc kubenswrapper[4730]: I0221 00:06:10.993090 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:10Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.007856 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.029686 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.058925 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.079111 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.108220 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.131973 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.147932 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.165129 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.180952 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.194542 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.210564 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.389681 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-28 14:50:41.24360627 +0000 UTC Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.691240 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-f7kbj" event={"ID":"a502eb4e-6f7f-4567-a7fd-bc15eb806845","Type":"ContainerStarted","Data":"4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe"} Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.696676 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76" exitCode=0 Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.696793 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76"} Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.705807 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d"} Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.705877 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061"} Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.705930 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260"} Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.714199 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.735372 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.757400 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.774131 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.791260 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.811107 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.827022 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.845501 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.861603 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.878838 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.894705 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.918077 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.936767 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.962439 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:11 crc kubenswrapper[4730]: I0221 00:06:11.984780 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.001437 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:11Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.001713 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.001816 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002009 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002035 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002052 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002043 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002089 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002107 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002121 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:16.002100995 +0000 UTC m=+28.292936735 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.002184 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:16.002163426 +0000 UTC m=+28.292999166 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.019464 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.041374 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.058375 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.075397 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.084727 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.096068 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.103436 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.103683 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:16.103646132 +0000 UTC m=+28.394481922 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.112504 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.126626 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.142473 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.158013 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.204651 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.204760 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.204812 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.204888 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:16.204869131 +0000 UTC m=+28.495704881 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.204894 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.204969 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:16.204955523 +0000 UTC m=+28.495791253 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.390458 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 18:15:18.510800553 +0000 UTC Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.434509 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.434509 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.434662 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.434787 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.434818 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:12 crc kubenswrapper[4730]: E0221 00:06:12.435299 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.717417 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c" exitCode=0 Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.717509 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c"} Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.720097 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa"} Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.737407 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.765713 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.780944 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.812780 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.831440 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.857370 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.874125 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.889486 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.908594 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.924412 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.936924 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.950503 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:12 crc kubenswrapper[4730]: I0221 00:06:12.966505 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:12Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.390863 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 05:56:44.21754493 +0000 UTC Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.731140 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d"} Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.735749 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229" exitCode=0 Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.735820 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229"} Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.759218 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.785141 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.805209 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.822512 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.846467 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.868868 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.890834 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.905048 4730 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.907392 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.909669 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.909729 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.909751 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.909948 4730 kubelet_node_status.go:76] "Attempting to register node" node="crc" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.922305 4730 kubelet_node_status.go:115] "Node was previously registered" node="crc" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.922589 4730 kubelet_node_status.go:79] "Successfully registered node" node="crc" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.924504 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.924574 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.924594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.924623 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.924642 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:13Z","lastTransitionTime":"2026-02-21T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.927444 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: E0221 00:06:13.943256 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948077 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948192 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948259 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948272 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948293 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.948308 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:13Z","lastTransitionTime":"2026-02-21T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:13 crc kubenswrapper[4730]: E0221 00:06:13.961015 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.968493 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.968555 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.968793 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.968831 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.968842 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:13Z","lastTransitionTime":"2026-02-21T00:06:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:13 crc kubenswrapper[4730]: I0221 00:06:13.974358 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:13 crc kubenswrapper[4730]: E0221 00:06:13.995222 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:13Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.004467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.004550 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.004570 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.004631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.004651 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.024329 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.074098 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.074373 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.078797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.078925 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.078956 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.078978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.079026 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.093606 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.093888 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.095985 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.096030 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.096049 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.096074 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.096093 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.199181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.199247 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.199264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.199289 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.199308 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.303309 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.303542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.303595 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.303623 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.303647 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.391779 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 16:14:33.779225233 +0000 UTC Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.407487 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.407549 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.407564 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.407588 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.407608 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.434722 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.435350 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.435560 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.435598 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.436118 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:14 crc kubenswrapper[4730]: E0221 00:06:14.437043 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.495344 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.503128 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.507811 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.510966 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.511632 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.511667 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.511691 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.511711 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.514450 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.533346 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.551892 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.576722 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.606169 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.614637 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.614684 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.614698 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.614718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.614733 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.634161 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.662220 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.683545 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.700631 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.717185 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.718391 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.718479 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.718509 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.718547 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.718572 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.737873 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.747123 4730 generic.go:334] "Generic (PLEG): container finished" podID="638e980f-5038-4a5d-85d2-4aae577964af" containerID="c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7" exitCode=0 Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.747218 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerDied","Data":"c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.762872 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.789971 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.815121 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.821473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.821577 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.821596 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.821617 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.821629 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.851858 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.871265 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.892049 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.909496 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.925730 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.925767 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.925780 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.925798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.925812 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:14Z","lastTransitionTime":"2026-02-21T00:06:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.931071 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.945338 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.960479 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.983024 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:14 crc kubenswrapper[4730]: I0221 00:06:14.999304 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:14Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.016177 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.029952 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.030377 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.030398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.030415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.030444 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.031361 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.048642 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.072919 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.133494 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.133584 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.133608 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.133641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.133665 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.237352 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.237424 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.237442 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.237471 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.237492 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.341088 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.341156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.341179 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.341210 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.341229 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.393331 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 17:20:05.130206408 +0000 UTC Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.444307 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.444896 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.444959 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.444989 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.445014 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.548535 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.548592 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.548611 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.548636 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.548655 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.652012 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.652066 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.652085 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.652107 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.652122 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.755780 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.755857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.755887 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.755954 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.755983 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.761454 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" event={"ID":"638e980f-5038-4a5d-85d2-4aae577964af","Type":"ContainerStarted","Data":"571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.769621 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.769977 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.770035 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.784031 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.796708 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.801275 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.804171 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.824395 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.842685 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.858452 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.858601 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.858660 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.858733 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.858842 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.861600 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.881243 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.898282 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.915847 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.930610 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.945970 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.961773 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.961829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.961846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.961868 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.961886 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:15Z","lastTransitionTime":"2026-02-21T00:06:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.965674 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:15 crc kubenswrapper[4730]: I0221 00:06:15.990529 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.010997 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.032767 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.048550 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.048606 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048761 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048779 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048795 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048857 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.048840482 +0000 UTC m=+36.339676222 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048869 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048927 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.048948 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.049030 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.049004237 +0000 UTC m=+36.339840027 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.052151 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.064700 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.064743 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.064760 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.064786 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.064806 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.066853 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.081055 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.100439 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.124834 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.149182 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.149513 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.149466244 +0000 UTC m=+36.440301974 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167295 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167577 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167670 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.167688 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.188726 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.207030 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.224156 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.237747 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.247391 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.251823 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.251949 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.252114 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.252093072 +0000 UTC m=+36.542928802 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.252141 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.252228 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.252334 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.252314788 +0000 UTC m=+36.543150518 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.260295 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.270555 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.270607 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.270616 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.270632 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.270642 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.275229 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.289610 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:16Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.374134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.374189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.374206 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.374227 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.374239 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.393570 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-13 00:48:09.130887264 +0000 UTC Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.434305 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.434353 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.434311 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.434447 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.434528 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:16 crc kubenswrapper[4730]: E0221 00:06:16.434682 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.477156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.477198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.477211 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.477230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.477240 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.580434 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.580476 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.580486 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.580503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.580513 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.684405 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.684446 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.684458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.684477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.684490 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.774071 4730 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.787247 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.787292 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.787514 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.787538 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.787551 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.891755 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.891824 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.891847 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.891876 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.891936 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.995374 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.995440 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.995455 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.995482 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:16 crc kubenswrapper[4730]: I0221 00:06:16.995498 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:16Z","lastTransitionTime":"2026-02-21T00:06:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.099260 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.099332 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.099350 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.099377 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.099394 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.204833 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.204880 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.204891 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.204929 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.204942 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.307940 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.308010 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.308028 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.308060 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.308078 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.394193 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-07 12:10:58.944293051 +0000 UTC Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.410574 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.410697 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.410718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.410746 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.410766 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.513738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.513820 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.513847 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.513886 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.513950 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.616840 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.616968 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.616997 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.617031 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.617056 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.720744 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.721249 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.721266 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.721292 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.721315 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.777824 4730 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.825774 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.826284 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.826505 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.826747 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.826961 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.930351 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.930414 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.930432 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.930450 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:17 crc kubenswrapper[4730]: I0221 00:06:17.930462 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:17Z","lastTransitionTime":"2026-02-21T00:06:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.034407 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.034446 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.034457 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.034473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.034482 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.137783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.137832 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.137841 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.137858 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.137870 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.241511 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.241584 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.241599 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.241620 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.241635 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.345143 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.345203 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.345215 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.345235 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.345248 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.395120 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-29 01:57:01.089751692 +0000 UTC Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.435095 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.435219 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:18 crc kubenswrapper[4730]: E0221 00:06:18.435691 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.435250 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:18 crc kubenswrapper[4730]: E0221 00:06:18.436240 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:18 crc kubenswrapper[4730]: E0221 00:06:18.435956 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.448197 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.448241 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.448254 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.448274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.448287 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.457141 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.468960 4730 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.483852 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.498375 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.517531 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.546620 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.551488 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.551537 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.551548 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.551565 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.551577 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.565967 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.588512 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.607556 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.621025 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.637312 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.653811 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.653856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.653868 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.653887 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.653917 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.662111 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.684890 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.700060 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.716251 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:18Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.756755 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.756808 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.756822 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.756839 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.756849 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.860114 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.860164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.860173 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.860191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.860201 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.963679 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.963730 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.963740 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.963780 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:18 crc kubenswrapper[4730]: I0221 00:06:18.963792 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:18Z","lastTransitionTime":"2026-02-21T00:06:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.068655 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.068728 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.068754 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.068780 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.068798 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.178011 4730 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.179598 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.179669 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.179689 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.179722 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.179745 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.283175 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.283239 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.283251 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.283273 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.283286 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.351767 4730 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.386554 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.386636 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.386656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.386686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.386706 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.396202 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 14:25:53.935178417 +0000 UTC Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.491081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.491135 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.491147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.491168 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.491183 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.594612 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.594656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.594665 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.594686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.594697 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.697583 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.697627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.697636 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.697656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.697668 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.787324 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/0.log" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.790745 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004" exitCode=1 Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.790811 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.791988 4730 scope.go:117] "RemoveContainer" containerID="9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.799448 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.799488 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.799501 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.799521 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.799535 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.813542 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.835307 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.858092 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.877822 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.898367 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.904541 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.904590 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.904603 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.904626 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.904641 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:19Z","lastTransitionTime":"2026-02-21T00:06:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.917058 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.935055 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.951121 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.967477 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:19 crc kubenswrapper[4730]: I0221 00:06:19.982993 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:19Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.005266 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.007039 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.007094 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.007106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.007122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.007134 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.026961 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.042362 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.057804 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.110751 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.110797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.110807 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.110823 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.110833 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.213992 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.214038 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.214053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.214072 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.214086 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.316805 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.316864 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.316882 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.316930 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.316944 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.397209 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-04 16:38:44.760137505 +0000 UTC Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.419262 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.419295 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.419305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.419323 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.419338 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.493604 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.493676 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:20 crc kubenswrapper[4730]: E0221 00:06:20.493790 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.493990 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:20 crc kubenswrapper[4730]: E0221 00:06:20.494106 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:20 crc kubenswrapper[4730]: E0221 00:06:20.494171 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.523097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.523144 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.523155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.523177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.523191 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.626468 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.626757 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.626831 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.626935 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.627009 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.723358 4730 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.730093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.730142 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.730155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.730174 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.730186 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.796185 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/0.log" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.798686 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.799010 4730 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.822506 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.832711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.832746 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.832755 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.832773 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.832784 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.839068 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.853964 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.867193 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.876740 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.891775 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.909077 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.923797 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.935198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.935258 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.935278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.935303 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.935322 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:20Z","lastTransitionTime":"2026-02-21T00:06:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.940132 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.952692 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.965797 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.979755 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:20 crc kubenswrapper[4730]: I0221 00:06:20.994202 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:20Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.006581 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.037750 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.037819 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.037831 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.037854 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.037869 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.140660 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.140740 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.140764 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.140791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.140812 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.243985 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.244056 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.244075 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.244102 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.244122 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.347158 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.347211 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.347223 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.347247 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.347258 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.398057 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 10:08:10.905846944 +0000 UTC Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.450947 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.451027 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.451037 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.451060 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.451072 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.554542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.554599 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.554627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.554649 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.554663 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.657600 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.657641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.657656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.657690 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.657711 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.761386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.761453 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.761467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.761491 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.761506 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.806404 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/1.log" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.807547 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/0.log" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.811828 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b" exitCode=1 Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.811883 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.812021 4730 scope.go:117] "RemoveContainer" containerID="9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.813724 4730 scope.go:117] "RemoveContainer" containerID="6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b" Feb 21 00:06:21 crc kubenswrapper[4730]: E0221 00:06:21.814131 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.838437 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.864857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.864952 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.864977 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.865009 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.865029 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.865672 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.885545 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.902278 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.919544 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.939869 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.962154 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.968738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.968784 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.968801 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.968825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.968842 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:21Z","lastTransitionTime":"2026-02-21T00:06:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:21 crc kubenswrapper[4730]: I0221 00:06:21.982818 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:21Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.003613 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.019471 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.050419 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.073158 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.073236 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.073275 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.073308 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.073329 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.076478 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.099820 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.121584 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx"] Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.124393 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.124947 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.126410 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.129614 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.152212 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.171101 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.176044 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.178021 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.178106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.178134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.178150 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.191228 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.213894 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.224410 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgj7l\" (UniqueName: \"kubernetes.io/projected/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-kube-api-access-tgj7l\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.224532 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.224557 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.224623 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.234119 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.249688 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.271887 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.281870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.281961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.281974 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.281995 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.282013 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.295692 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.315707 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.325956 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.326005 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgj7l\" (UniqueName: \"kubernetes.io/projected/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-kube-api-access-tgj7l\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.326060 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.326087 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.327537 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-env-overrides\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.327689 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.333579 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.336073 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.346703 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgj7l\" (UniqueName: \"kubernetes.io/projected/f8a36a70-028a-4dca-87d5-b4d1a8b1d048-kube-api-access-tgj7l\") pod \"ovnkube-control-plane-749d76644c-k28mx\" (UID: \"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.350581 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.364319 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.380752 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.386039 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.386118 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.386136 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.386208 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.386224 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.396437 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.398656 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 15:05:21.009091027 +0000 UTC Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.411540 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.434169 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.434207 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.434297 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:22 crc kubenswrapper[4730]: E0221 00:06:22.434355 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:22 crc kubenswrapper[4730]: E0221 00:06:22.434437 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:22 crc kubenswrapper[4730]: E0221 00:06:22.434562 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.440770 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" Feb 21 00:06:22 crc kubenswrapper[4730]: W0221 00:06:22.457567 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8a36a70_028a_4dca_87d5_b4d1a8b1d048.slice/crio-5dc5934e5c7a116599722367fe65a8362cfac2cb01f4d570ed078b37c6b8645c WatchSource:0}: Error finding container 5dc5934e5c7a116599722367fe65a8362cfac2cb01f4d570ed078b37c6b8645c: Status 404 returned error can't find the container with id 5dc5934e5c7a116599722367fe65a8362cfac2cb01f4d570ed078b37c6b8645c Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.490573 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.490631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.490643 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.490661 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.490673 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.594518 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.594577 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.594600 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.594625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.594640 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.697119 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.697160 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.697169 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.697185 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.697196 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.803079 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.803119 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.803129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.803146 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.803159 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.816979 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" event={"ID":"f8a36a70-028a-4dca-87d5-b4d1a8b1d048","Type":"ContainerStarted","Data":"3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.817064 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" event={"ID":"f8a36a70-028a-4dca-87d5-b4d1a8b1d048","Type":"ContainerStarted","Data":"095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.817081 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" event={"ID":"f8a36a70-028a-4dca-87d5-b4d1a8b1d048","Type":"ContainerStarted","Data":"5dc5934e5c7a116599722367fe65a8362cfac2cb01f4d570ed078b37c6b8645c"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.818999 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/1.log" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.836810 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.854197 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.875082 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.877825 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-krrw8"] Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.878418 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:22 crc kubenswrapper[4730]: E0221 00:06:22.878482 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.889954 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906075 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906121 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906138 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906158 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906171 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:22Z","lastTransitionTime":"2026-02-21T00:06:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.906283 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.921982 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.943948 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.971514 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.985420 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:22 crc kubenswrapper[4730]: I0221 00:06:22.998735 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:22Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.008740 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.008800 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.008814 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.008839 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.008854 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.011069 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.024853 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.034444 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.034530 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjwnh\" (UniqueName: \"kubernetes.io/projected/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-kube-api-access-rjwnh\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.042371 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.057733 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.090169 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.112321 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.112388 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.112406 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.112433 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.112452 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.119019 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9457e5b452c988ca0f8442e8137c4b2a16a9c68a1034ca3de4192b49c1f8f004\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:18Z\\\",\\\"message\\\":\\\" 6039 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0221 00:06:18.751111 6039 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0221 00:06:18.751190 6039 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0221 00:06:18.751206 6039 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0221 00:06:18.751221 6039 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0221 00:06:18.751227 6039 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0221 00:06:18.751249 6039 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0221 00:06:18.751308 6039 handler.go:208] Removed *v1.Node event handler 2\\\\nI0221 00:06:18.751306 6039 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0221 00:06:18.751334 6039 handler.go:208] Removed *v1.Node event handler 7\\\\nI0221 00:06:18.751348 6039 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0221 00:06:18.751374 6039 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0221 00:06:18.751377 6039 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0221 00:06:18.751422 6039 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0221 00:06:18.751453 6039 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0221 00:06:18.751445 6039 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.135254 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjwnh\" (UniqueName: \"kubernetes.io/projected/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-kube-api-access-rjwnh\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.135430 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: E0221 00:06:23.135638 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:23 crc kubenswrapper[4730]: E0221 00:06:23.135724 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:23.635700359 +0000 UTC m=+35.926536129 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.174704 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.186266 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjwnh\" (UniqueName: \"kubernetes.io/projected/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-kube-api-access-rjwnh\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.191075 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.212003 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.216044 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.216101 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.216122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.216149 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.216169 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.237190 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.252383 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.267753 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.285808 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.305335 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.318974 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.319042 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.319053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.319078 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.319093 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.323688 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.336605 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.354258 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.372874 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.391399 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.399070 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 18:59:55.451273726 +0000 UTC Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.408543 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.415203 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.416492 4730 scope.go:117] "RemoveContainer" containerID="6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b" Feb 21 00:06:23 crc kubenswrapper[4730]: E0221 00:06:23.416850 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.422637 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.422682 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.422699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.422725 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.422744 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.428788 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.447654 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.466186 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.484581 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.499875 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.519653 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.525631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.525680 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.525693 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.525713 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.525730 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.553874 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.587314 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.605515 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.621939 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.628088 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.628134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.628148 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.628165 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.628178 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.639582 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.642228 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:23 crc kubenswrapper[4730]: E0221 00:06:23.642351 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:23 crc kubenswrapper[4730]: E0221 00:06:23.642398 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:24.642385357 +0000 UTC m=+36.933221077 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.662138 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.674794 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.687802 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.704031 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.717554 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.729043 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:23Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.731038 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.731073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.731085 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.731106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.731124 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.833994 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.834046 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.834057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.834076 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.834090 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.937746 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.937833 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.937853 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.937879 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:23 crc kubenswrapper[4730]: I0221 00:06:23.937896 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:23Z","lastTransitionTime":"2026-02-21T00:06:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.040398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.040458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.040469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.040487 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.040498 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.143428 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.143520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.143534 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.143556 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.143568 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.147094 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.147152 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147316 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147339 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147345 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147361 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147369 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147379 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147461 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:40.14743559 +0000 UTC m=+52.438271500 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.147483 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:40.147474981 +0000 UTC m=+52.438310711 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.199007 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.199067 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.199077 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.199132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.199146 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.216774 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:24Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.222578 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.222635 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.222645 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.222666 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.222680 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.242215 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:24Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.247839 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.247896 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.247931 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.247951 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.247965 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.248022 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.248353 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:06:40.248299257 +0000 UTC m=+52.539135017 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.278355 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:24Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.284607 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.284654 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.284669 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.284691 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.284705 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.305497 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:24Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.311706 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.311746 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.311760 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.311776 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.311790 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.335647 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:24Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:24Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.336013 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.338861 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.338961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.338983 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.339011 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.339032 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.349551 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.349645 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.349712 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.349799 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:40.349777873 +0000 UTC m=+52.640613613 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.351018 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.351114 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:40.35109326 +0000 UTC m=+52.641929030 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.399602 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 08:20:05.932489376 +0000 UTC Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.434082 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.434150 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.434151 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.434157 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.434298 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.434482 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.434617 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.434745 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.441931 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.441982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.442001 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.442023 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.442042 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.545298 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.545357 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.545369 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.545387 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.545400 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.648738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.648788 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.648798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.648815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.648825 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.653519 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.653706 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: E0221 00:06:24.653812 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:26.653787037 +0000 UTC m=+38.944622797 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.752817 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.752871 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.752883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.752936 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.752954 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.855806 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.855884 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.855938 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.856057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.856164 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.958469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.958526 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.958543 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.958567 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:24 crc kubenswrapper[4730]: I0221 00:06:24.958584 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:24Z","lastTransitionTime":"2026-02-21T00:06:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.061661 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.061722 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.061741 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.061770 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.061790 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.165088 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.165149 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.165166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.165190 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.165213 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.268817 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.268933 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.268955 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.268981 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.269000 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.372781 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.372847 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.372860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.372882 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.372896 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.400057 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-18 01:53:24.000981996 +0000 UTC Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.476477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.476527 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.476539 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.476563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.476579 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.579644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.579718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.579736 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.579763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.579785 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.683200 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.683261 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.683277 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.683307 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.683324 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.786651 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.786706 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.786718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.786734 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.786745 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.890068 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.890742 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.890769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.890804 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.890829 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.994677 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.994756 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.994779 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.994813 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:25 crc kubenswrapper[4730]: I0221 00:06:25.994839 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:25Z","lastTransitionTime":"2026-02-21T00:06:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.098741 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.098856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.098885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.098948 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.098970 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.202468 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.202532 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.202545 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.202568 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.202581 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.305516 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.305565 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.305575 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.305591 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.305601 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.400772 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 22:00:31.417904372 +0000 UTC Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.408236 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.408303 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.408314 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.408331 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.408344 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.434015 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.434015 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.434173 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.434244 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.434358 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.434496 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.434601 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.434688 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.511350 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.511448 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.511471 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.511504 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.511529 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.615267 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.615326 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.615337 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.615361 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.615375 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.678178 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.678383 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:26 crc kubenswrapper[4730]: E0221 00:06:26.678505 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:30.678464053 +0000 UTC m=+42.969299823 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.718734 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.718797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.718809 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.718827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.718838 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.822513 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.822583 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.822608 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.822642 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.822666 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.925075 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.925171 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.925198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.925231 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:26 crc kubenswrapper[4730]: I0221 00:06:26.925257 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:26Z","lastTransitionTime":"2026-02-21T00:06:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.028590 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.028662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.028684 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.028712 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.028730 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.132024 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.132089 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.132111 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.132158 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.132189 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.235020 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.235091 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.235111 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.235140 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.235159 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.254345 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.277246 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.294509 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.310867 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.332029 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.337978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.338059 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.338081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.338109 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.338130 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.352296 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.394645 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.401756 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-21 11:23:07.985033159 +0000 UTC Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.436455 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.441419 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.441473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.441486 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.441510 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.441527 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.454707 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.471284 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.488378 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.503648 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.518643 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.538940 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.544038 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.544097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.544110 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.544141 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.544158 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.560480 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.578297 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.594387 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:27Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.646768 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.646837 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.646856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.646883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.646952 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.750464 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.750555 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.750579 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.750610 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.750629 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.854366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.854460 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.854479 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.854508 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.854526 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.963338 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.963415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.963436 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.963464 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:27 crc kubenswrapper[4730]: I0221 00:06:27.963550 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:27Z","lastTransitionTime":"2026-02-21T00:06:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.066759 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.066813 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.066826 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.066846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.066861 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.170988 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.171049 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.171067 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.171091 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.171109 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.274458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.274522 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.274537 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.274556 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.274573 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.377797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.377856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.377869 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.377888 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.377924 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.402691 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 20:44:12.791581619 +0000 UTC Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.434238 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.434329 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.434466 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:28 crc kubenswrapper[4730]: E0221 00:06:28.434619 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.434749 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:28 crc kubenswrapper[4730]: E0221 00:06:28.434792 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:28 crc kubenswrapper[4730]: E0221 00:06:28.434995 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:28 crc kubenswrapper[4730]: E0221 00:06:28.436053 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.454623 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.469971 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.481023 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.481096 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.481115 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.481141 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.481160 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.484154 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.504561 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.519783 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.540232 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.557475 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.579669 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.587732 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.587820 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.587845 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.587878 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.587933 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.595433 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.613674 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.633089 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.647611 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.672449 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.690805 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.691461 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.691564 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.691645 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.691711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.691771 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.710790 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.733971 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:28Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.794693 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.795122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.795241 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.795357 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.795468 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.898222 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.898284 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.898301 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.898326 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:28 crc kubenswrapper[4730]: I0221 00:06:28.898345 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:28Z","lastTransitionTime":"2026-02-21T00:06:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.002086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.002148 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.002167 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.002193 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.002211 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.104443 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.104579 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.104613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.104645 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.104671 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.207478 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.207539 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.207550 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.207571 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.207585 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.309790 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.309835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.309846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.309864 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.309877 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.402952 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-08 21:09:36.933856028 +0000 UTC Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.412747 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.412804 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.412820 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.412849 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.412869 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.517191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.517243 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.517260 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.517284 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.517303 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.620212 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.620274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.620291 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.620315 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.620333 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.723851 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.723941 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.723962 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.723988 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.724009 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.826276 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.826330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.826389 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.826420 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.826436 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.930465 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.930546 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.930571 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.930600 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:29 crc kubenswrapper[4730]: I0221 00:06:29.930620 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:29Z","lastTransitionTime":"2026-02-21T00:06:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.033633 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.033699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.033718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.033741 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.033759 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.136990 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.137048 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.137061 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.137082 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.137122 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.240722 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.240783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.240801 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.240824 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.240843 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.343679 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.343743 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.343760 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.343783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.343799 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.403533 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-08 08:51:52.027486612 +0000 UTC Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.434714 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.434833 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.435686 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.434931 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.435817 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.435031 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.434847 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.435980 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.445999 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.446053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.446069 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.446093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.446111 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.550325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.550404 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.550425 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.550454 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.550476 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.653537 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.653594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.653614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.653641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.653662 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.725480 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:30 crc kubenswrapper[4730]: E0221 00:06:30.725586 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:38.725562151 +0000 UTC m=+51.016397911 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.725300 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.757128 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.757194 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.757208 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.757238 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.757254 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.860059 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.860115 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.860130 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.860147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.860160 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.963535 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.963586 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.963602 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.963622 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:30 crc kubenswrapper[4730]: I0221 00:06:30.963634 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:30Z","lastTransitionTime":"2026-02-21T00:06:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.067208 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.067280 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.067301 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.067330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.067356 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.171474 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.171563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.171586 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.171620 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.171640 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.275002 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.275074 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.275098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.275132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.275156 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.378329 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.378447 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.378478 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.378519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.378544 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.404733 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-30 12:49:44.641519595 +0000 UTC Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.482073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.482165 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.482189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.482224 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.482248 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.586201 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.586289 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.586310 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.586340 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.586362 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.689667 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.689710 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.689723 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.689742 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.689754 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.793017 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.793095 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.793113 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.793145 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.793167 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.896190 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.896233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.896246 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.896263 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.896272 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.999105 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.999179 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.999198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.999230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:31 crc kubenswrapper[4730]: I0221 00:06:31.999251 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:31Z","lastTransitionTime":"2026-02-21T00:06:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.101892 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.101991 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.102002 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.102021 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.102033 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.205240 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.205282 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.205291 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.205308 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.205318 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.308150 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.308253 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.308273 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.308349 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.308368 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.404947 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-12 13:35:53.796012541 +0000 UTC Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.411012 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.411060 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.411073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.411090 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.411101 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.434671 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.434722 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.434699 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.434670 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:32 crc kubenswrapper[4730]: E0221 00:06:32.434878 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:32 crc kubenswrapper[4730]: E0221 00:06:32.435048 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:32 crc kubenswrapper[4730]: E0221 00:06:32.435274 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:32 crc kubenswrapper[4730]: E0221 00:06:32.435358 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.514477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.514539 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.514556 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.514580 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.514598 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.618216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.618330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.618347 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.618366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.618421 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.721648 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.721707 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.721723 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.721747 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.721764 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.825188 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.825240 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.825254 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.825274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.825288 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.927962 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.928012 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.928025 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.928043 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:32 crc kubenswrapper[4730]: I0221 00:06:32.928054 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:32Z","lastTransitionTime":"2026-02-21T00:06:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.031587 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.031651 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.031844 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.031866 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.031884 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.134958 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.135017 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.135027 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.135045 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.135063 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.258657 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.258749 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.258763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.258810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.258828 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.363444 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.363505 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.363518 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.363541 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.363556 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.405619 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-15 00:21:49.855064033 +0000 UTC Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.466650 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.466719 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.466738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.466764 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.466788 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.570607 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.570691 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.570715 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.570743 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.570762 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.674388 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.674456 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.674478 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.674504 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.674523 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.777540 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.777627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.777653 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.777686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.777708 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.881044 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.881097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.881108 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.881127 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.881140 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.984706 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.984790 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.984817 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.984846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:33 crc kubenswrapper[4730]: I0221 00:06:33.984864 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:33Z","lastTransitionTime":"2026-02-21T00:06:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.088399 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.088460 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.088477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.088504 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.088522 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.191981 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.192057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.192081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.192118 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.192148 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.295400 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.295477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.295503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.295539 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.295563 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.351690 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.351760 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.351847 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.351880 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.351945 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.374292 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.380355 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.380413 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.380430 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.380451 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.380475 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.400748 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.405748 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-29 20:42:03.332734063 +0000 UTC Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.406022 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.406059 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.406073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.406087 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.406098 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.425329 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.431230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.431303 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.431322 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.431347 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.431366 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.433918 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.433966 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.434070 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.434243 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.434321 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.434477 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.434582 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.434676 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.450395 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.456200 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.456245 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.456258 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.456278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.456291 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.476182 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:34Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:34Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:34 crc kubenswrapper[4730]: E0221 00:06:34.476407 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.479177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.479218 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.479234 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.479261 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.479280 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.582153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.582186 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.582195 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.582210 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.582218 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.686474 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.686530 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.686548 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.686572 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.686591 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.789324 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.789398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.789421 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.789478 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.789500 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.892517 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.892585 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.892605 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.892631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.892652 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.996384 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.996456 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.996479 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.996506 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:34 crc kubenswrapper[4730]: I0221 00:06:34.996525 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:34Z","lastTransitionTime":"2026-02-21T00:06:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.100128 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.100197 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.100214 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.100239 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.100274 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.204422 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.204477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.204492 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.204516 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.204536 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.307825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.307990 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.308019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.308045 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.308063 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.406898 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 21:40:25.450651248 +0000 UTC Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.411829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.411898 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.411949 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.411976 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.411997 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.515162 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.515255 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.515275 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.515307 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.515326 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.619185 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.619261 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.619278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.619305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.619325 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.723231 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.723320 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.723339 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.723366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.723384 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.826768 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.826827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.826846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.826870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.826889 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.929556 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.929623 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.929641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.929669 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:35 crc kubenswrapper[4730]: I0221 00:06:35.929691 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:35Z","lastTransitionTime":"2026-02-21T00:06:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.033180 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.033264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.033284 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.033314 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.033348 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.141472 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.141538 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.141561 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.141596 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.141620 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.245286 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.245370 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.245397 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.245432 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.245456 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.349055 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.349119 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.349137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.349163 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.349184 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.407216 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-15 00:43:00.678848368 +0000 UTC Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.435148 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.435255 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.435299 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.435364 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:36 crc kubenswrapper[4730]: E0221 00:06:36.435753 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:36 crc kubenswrapper[4730]: E0221 00:06:36.435964 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:36 crc kubenswrapper[4730]: E0221 00:06:36.436195 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.436282 4730 scope.go:117] "RemoveContainer" containerID="6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b" Feb 21 00:06:36 crc kubenswrapper[4730]: E0221 00:06:36.436479 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.452509 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.452567 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.452586 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.452612 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.452632 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.556072 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.556143 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.556164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.556193 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.556213 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.659644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.660136 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.660153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.660177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.660195 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.763741 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.763808 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.763832 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.763863 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.763887 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.870793 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.870876 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.870898 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.870958 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.870977 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.885647 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/1.log" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.890592 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.891526 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.918825 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.947507 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.970596 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.975029 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.975079 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.975095 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.975124 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.975139 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:36Z","lastTransitionTime":"2026-02-21T00:06:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:36 crc kubenswrapper[4730]: I0221 00:06:36.995034 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:36Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.020593 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.040387 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.058360 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.072695 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.077099 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.077141 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.077159 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.077182 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.077201 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.089159 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.102749 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.117127 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.141115 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.162945 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.180520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.180585 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.180605 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.180631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.180648 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.193791 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.219842 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.238054 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.282514 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.282558 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.282567 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.282584 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.282596 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.385536 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.385594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.385606 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.385629 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.385647 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.408106 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-05 05:31:54.229345132 +0000 UTC Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.488739 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.488825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.488843 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.488869 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.488889 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.592218 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.592282 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.592302 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.592330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.592350 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.695213 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.695297 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.695324 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.695356 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.695379 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.798940 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.799018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.799044 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.799073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.799098 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.897872 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/2.log" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.899088 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/1.log" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.901264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.901323 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.901346 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.901375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.901395 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:37Z","lastTransitionTime":"2026-02-21T00:06:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.906531 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" exitCode=1 Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.906604 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01"} Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.906992 4730 scope.go:117] "RemoveContainer" containerID="6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.907725 4730 scope.go:117] "RemoveContainer" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" Feb 21 00:06:37 crc kubenswrapper[4730]: E0221 00:06:37.908053 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.945116 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.971199 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:37 crc kubenswrapper[4730]: I0221 00:06:37.990360 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.004532 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.004570 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.004588 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.004613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.004631 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.012200 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.033198 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.060196 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.082805 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.102276 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.108696 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.108754 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.108774 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.108803 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.108857 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.123223 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.137800 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.147163 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.167057 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.184497 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.201729 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.212625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.212671 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.212680 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.212695 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.212704 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.218764 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.233622 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.316484 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.316557 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.316575 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.316603 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.316623 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.408995 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 01:23:38.946109835 +0000 UTC Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.420587 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.420697 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.420724 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.420757 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.420782 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.435132 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.435218 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.435279 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.435166 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.435416 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.435564 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.435795 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.435964 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.457545 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.476882 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.502830 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.524476 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.524551 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.524570 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.524599 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.524651 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.528862 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.551369 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.575536 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.597133 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.617404 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.627830 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.628531 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.628596 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.628610 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.628632 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.628648 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.641238 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.653844 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.692035 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6e7d6b276a924ed2a39e3f5ffac7c5a23ee69c1fc8d80c1cc8e63527f16d9b7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:21Z\\\",\\\"message\\\":\\\"t:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.150\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI0221 00:06:20.970524 6175 services_controller.go:452] Built service openshift-authentication-operator/metrics per-node LB for network=default: []services.LB{}\\\\nI0221 00:06:20.970531 6175 services_controller.go:453] Built service openshift-authentication-operator/metrics template LB for network=default: []services.LB{}\\\\nF0221 00:06:20.970533 6175 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:19Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.709278 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.726136 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.730806 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.731067 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.731150 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:06:54.731128304 +0000 UTC m=+67.021964044 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.731614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.731653 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.731665 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.731686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.731702 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.741775 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.757819 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.836338 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.836832 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.836950 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.837056 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.837151 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.913622 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/2.log" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.919758 4730 scope.go:117] "RemoveContainer" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" Feb 21 00:06:38 crc kubenswrapper[4730]: E0221 00:06:38.920415 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.938844 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.941542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.941630 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.941644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.941668 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.941691 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:38Z","lastTransitionTime":"2026-02-21T00:06:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.963954 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.982816 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:38 crc kubenswrapper[4730]: I0221 00:06:38.995598 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:38Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.013751 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.029586 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.045263 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.045331 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.045351 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.045381 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.045401 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.054224 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.078038 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.097545 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.112994 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.131377 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.148682 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.150729 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.150848 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.150875 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.150982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.151050 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.161508 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.177307 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.195018 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.216407 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.219397 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.230383 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.245336 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.255053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.255154 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.255177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.255207 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.255226 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.262402 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.286102 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.302773 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.318310 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.332588 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.346397 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.358178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.358260 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.358290 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.358323 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.358352 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.359501 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.370758 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.384602 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.407203 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.409261 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 06:06:46.024848154 +0000 UTC Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.425376 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.445086 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.462115 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.463035 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.463066 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.463102 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.463126 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.467831 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.481414 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.497459 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:39Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.566191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.566269 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.566282 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.566304 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.566316 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.669614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.670142 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.670380 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.670575 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.670755 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.773939 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.773999 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.774016 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.774041 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.774057 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.877220 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.877305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.877330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.877368 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.877390 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.980380 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.980776 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.980897 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.981089 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:39 crc kubenswrapper[4730]: I0221 00:06:39.981240 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:39Z","lastTransitionTime":"2026-02-21T00:06:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.084575 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.084941 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.085103 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.085263 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.085407 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.189036 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.189087 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.189103 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.189124 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.189136 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.245606 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.245690 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.245879 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.245957 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.245981 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.245979 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.246038 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.246065 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.246080 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:07:12.246048055 +0000 UTC m=+84.536883825 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.246171 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:07:12.246132948 +0000 UTC m=+84.536968718 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.292093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.292378 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.292512 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.292677 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.292817 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.346831 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.347064 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:07:12.347032027 +0000 UTC m=+84.637867787 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.396486 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.396567 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.396595 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.396625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.396649 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.409617 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:21:45.853310796 +0000 UTC Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.434395 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.434440 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.434414 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.434414 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.434600 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.434690 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.434745 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.434817 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.448029 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.448109 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.448204 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.448230 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.448275 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:07:12.448255546 +0000 UTC m=+84.739091286 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: E0221 00:06:40.448293 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:07:12.448284726 +0000 UTC m=+84.739120476 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.499605 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.499670 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.499762 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.499798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.499825 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.603322 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.603398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.603432 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.603467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.603489 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.706795 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.706850 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.706858 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.706878 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.706893 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.826736 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.826806 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.826826 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.826852 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.826871 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.930532 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.930617 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.930642 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.930670 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:40 crc kubenswrapper[4730]: I0221 00:06:40.930688 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:40Z","lastTransitionTime":"2026-02-21T00:06:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.034535 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.034582 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.034594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.034611 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.034622 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.138340 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.138392 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.138403 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.138422 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.138433 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.241670 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.241749 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.241769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.241799 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.241821 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.345896 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.346012 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.346035 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.346061 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.346080 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.411144 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 12:22:59.048816523 +0000 UTC Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.448978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.449049 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.449068 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.449093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.449114 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.552034 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.552156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.552182 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.552216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.552237 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.655595 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.655643 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.655651 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.655666 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.655679 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.758615 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.758683 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.758703 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.758729 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.758750 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.862138 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.862216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.862237 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.862267 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.862288 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.965127 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.965198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.965217 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.965243 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:41 crc kubenswrapper[4730]: I0221 00:06:41.965261 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:41Z","lastTransitionTime":"2026-02-21T00:06:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.067869 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.067982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.068003 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.068029 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.068050 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.170797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.170845 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.170860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.170881 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.170948 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.273959 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.274040 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.274058 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.274087 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.274107 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.378078 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.378137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.378154 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.378178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.378198 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.411612 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 21:15:57.562691609 +0000 UTC Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.434598 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.434724 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.434809 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:42 crc kubenswrapper[4730]: E0221 00:06:42.434890 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:42 crc kubenswrapper[4730]: E0221 00:06:42.435021 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.435130 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:42 crc kubenswrapper[4730]: E0221 00:06:42.435285 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:42 crc kubenswrapper[4730]: E0221 00:06:42.435514 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.481375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.481445 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.481459 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.481477 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.481491 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.584493 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.584545 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.584561 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.584589 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.584607 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.687069 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.687129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.687142 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.687162 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.687175 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.790465 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.790530 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.790549 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.790577 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.790594 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.893855 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.894002 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.894032 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.894065 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.894086 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.996765 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.996815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.996831 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.996852 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:42 crc kubenswrapper[4730]: I0221 00:06:42.996870 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:42Z","lastTransitionTime":"2026-02-21T00:06:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.100325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.100439 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.100467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.100543 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.100568 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.203392 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.203450 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.203468 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.203490 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.203504 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.306620 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.306683 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.306700 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.306723 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.306742 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.410122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.410183 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.410202 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.410231 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.410251 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.411928 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 21:09:40.938421429 +0000 UTC Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.513990 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.514078 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.514097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.514126 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.514145 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.617582 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.617682 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.617710 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.617753 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.617780 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.721038 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.721112 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.721129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.721156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.721178 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.824474 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.824543 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.824563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.824587 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.824658 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.928452 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.928820 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.928872 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.928957 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:43 crc kubenswrapper[4730]: I0221 00:06:43.928987 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:43Z","lastTransitionTime":"2026-02-21T00:06:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.032015 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.032080 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.032097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.032125 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.032143 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.141589 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.141670 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.141689 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.141728 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.141747 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.244893 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.244962 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.244986 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.245000 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.245009 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.348485 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.348613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.348633 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.348660 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.348678 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.412321 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-11 20:33:19.875865712 +0000 UTC Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.434833 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.434852 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.434990 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.435059 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.435111 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.435192 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.435336 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.435583 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.451122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.451187 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.451208 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.451233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.451264 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.548150 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.548220 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.548239 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.548264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.548283 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.569619 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.574898 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.574981 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.574999 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.575024 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.575043 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.596174 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.601388 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.601446 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.601472 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.601501 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.601522 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.622349 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.627206 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.627250 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.627265 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.627286 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.627302 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.647616 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.652783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.652834 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.652856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.652883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.652943 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.673176 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:44Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:44Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:44 crc kubenswrapper[4730]: E0221 00:06:44.673392 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.675500 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.675545 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.675561 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.675583 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.675602 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.779137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.779207 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.779230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.779261 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.779283 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.883082 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.883147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.883173 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.883203 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.883227 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.985779 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.985835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.985851 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.985875 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:44 crc kubenswrapper[4730]: I0221 00:06:44.985893 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:44Z","lastTransitionTime":"2026-02-21T00:06:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.088876 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.088975 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.088996 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.089023 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.089046 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.192001 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.192062 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.192122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.192149 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.192196 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.296081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.296123 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.296136 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.296156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.296169 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.400662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.400772 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.400791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.400816 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.400834 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.413180 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 22:13:38.769882389 +0000 UTC Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.503968 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.504037 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.504056 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.504080 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.504097 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.607293 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.607399 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.607417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.607445 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.607468 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.711966 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.712034 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.712052 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.712079 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.712097 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.815664 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.815738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.815766 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.815801 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.815825 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.919583 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.919656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.919674 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.919711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:45 crc kubenswrapper[4730]: I0221 00:06:45.919744 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:45Z","lastTransitionTime":"2026-02-21T00:06:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.023378 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.023424 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.023434 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.023453 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.023464 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.128237 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.128310 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.128334 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.128366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.128389 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.232540 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.232614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.232635 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.232657 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.232670 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.335139 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.335191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.335203 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.335220 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.335233 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.413397 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 13:57:40.730017352 +0000 UTC Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.434885 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.435004 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.434894 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:46 crc kubenswrapper[4730]: E0221 00:06:46.435194 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.435223 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:46 crc kubenswrapper[4730]: E0221 00:06:46.435435 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:46 crc kubenswrapper[4730]: E0221 00:06:46.435713 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:46 crc kubenswrapper[4730]: E0221 00:06:46.435788 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.438312 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.438416 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.438481 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.438505 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.438571 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.542118 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.542189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.542213 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.542247 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.542268 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.646076 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.646138 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.646162 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.646192 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.646213 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.749775 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.749835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.749848 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.749868 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.749881 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.864453 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.864559 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.864584 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.864665 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.864727 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.968320 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.968368 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.968380 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.968398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:46 crc kubenswrapper[4730]: I0221 00:06:46.968410 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:46Z","lastTransitionTime":"2026-02-21T00:06:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.071841 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.071961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.071982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.072014 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.072034 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.175797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.175855 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.175864 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.175883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.175899 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.280540 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.280621 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.280647 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.280683 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.280711 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.384091 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.384155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.384173 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.384200 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.384221 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.413662 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 17:15:02.708537601 +0000 UTC Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.488223 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.488289 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.488306 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.488332 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.488349 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.592212 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.592303 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.592327 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.592405 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.592431 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.695716 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.695782 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.695794 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.695821 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.695836 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.799748 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.799829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.799850 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.799882 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.799949 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.903076 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.903164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.903180 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.903206 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:47 crc kubenswrapper[4730]: I0221 00:06:47.903227 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:47Z","lastTransitionTime":"2026-02-21T00:06:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.006683 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.006757 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.006781 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.006812 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.006833 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.109827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.109871 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.109880 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.109896 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.109931 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.213732 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.213792 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.213805 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.213825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.213840 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.317153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.317305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.317326 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.317353 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.317371 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.414782 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 12:48:32.596630555 +0000 UTC Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.419715 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.419773 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.419791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.419818 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.419871 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.437675 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:48 crc kubenswrapper[4730]: E0221 00:06:48.437967 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.437999 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.438058 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.438068 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:48 crc kubenswrapper[4730]: E0221 00:06:48.438170 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:48 crc kubenswrapper[4730]: E0221 00:06:48.438508 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:48 crc kubenswrapper[4730]: E0221 00:06:48.440285 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.460033 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.482951 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.496862 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.518660 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.524791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.524858 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.524878 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.524930 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.524954 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.536331 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.563416 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.601562 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.628142 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.628182 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.628194 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.628215 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.628228 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.630976 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.655242 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.668088 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.683739 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.698161 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.715744 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.732079 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.732140 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.732161 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.732189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.732211 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.733300 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.761082 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.788065 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.807107 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:48Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.835588 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.835625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.835638 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.835655 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.835670 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.940028 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.940084 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.940098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.940122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:48 crc kubenswrapper[4730]: I0221 00:06:48.940137 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:48Z","lastTransitionTime":"2026-02-21T00:06:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.043025 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.043108 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.043128 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.043167 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.043202 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.146982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.147063 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.147082 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.147111 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.147134 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.250716 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.250770 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.250788 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.250811 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.250826 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.354688 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.354751 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.354770 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.354799 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.354818 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.415423 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 07:06:45.553366323 +0000 UTC Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.458471 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.458547 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.458566 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.458594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.458617 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.562695 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.562753 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.562771 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.562799 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.562817 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.666803 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.666874 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.666926 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.666953 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.666973 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.771652 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.771699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.771711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.771733 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.771747 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.874430 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.874558 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.874585 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.874622 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.874646 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.977708 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.977769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.977783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.977811 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:49 crc kubenswrapper[4730]: I0221 00:06:49.977828 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:49Z","lastTransitionTime":"2026-02-21T00:06:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.081856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.081961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.081982 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.082013 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.082034 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.187377 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.187442 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.187475 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.187500 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.187518 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.290604 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.290685 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.290704 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.290732 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.290754 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.394827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.394879 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.394895 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.394945 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.394963 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.415648 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-25 00:45:05.920620777 +0000 UTC Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.434847 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.434987 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.434856 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.435128 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:50 crc kubenswrapper[4730]: E0221 00:06:50.435140 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:50 crc kubenswrapper[4730]: E0221 00:06:50.435495 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:50 crc kubenswrapper[4730]: E0221 00:06:50.435627 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:50 crc kubenswrapper[4730]: E0221 00:06:50.435734 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.498860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.498994 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.499019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.499054 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.499076 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.603018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.603096 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.603116 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.603147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.603168 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.706686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.706763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.706782 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.706810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.706831 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.810236 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.810285 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.810298 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.810319 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.810336 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.912822 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.912857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.912865 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.912877 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:50 crc kubenswrapper[4730]: I0221 00:06:50.912887 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:50Z","lastTransitionTime":"2026-02-21T00:06:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.015244 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.015325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.015344 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.015369 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.015387 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.117542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.117596 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.117607 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.117622 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.117638 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.221225 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.221278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.221296 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.221318 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.221335 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.325038 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.325103 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.325120 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.325146 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.325165 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.415963 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-02 01:57:57.804081421 +0000 UTC Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.431965 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.432030 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.432054 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.432086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.432106 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.535995 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.536070 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.536090 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.536117 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.536138 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.639955 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.640013 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.640032 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.640056 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.640074 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.743358 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.743417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.743438 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.743464 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.743481 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.847345 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.847448 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.847473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.847503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.847525 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.951329 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.951407 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.951419 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.951438 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:51 crc kubenswrapper[4730]: I0221 00:06:51.951450 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:51Z","lastTransitionTime":"2026-02-21T00:06:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.054687 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.054758 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.054770 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.054801 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.054818 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.158251 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.158325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.158339 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.158361 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.158375 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.261981 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.262047 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.262074 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.262108 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.262165 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.366271 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.366339 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.366359 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.366383 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.366409 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.416307 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 22:07:35.346144896 +0000 UTC Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.434225 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.434266 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.434334 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:52 crc kubenswrapper[4730]: E0221 00:06:52.434460 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.434574 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:52 crc kubenswrapper[4730]: E0221 00:06:52.434710 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:52 crc kubenswrapper[4730]: E0221 00:06:52.434851 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:52 crc kubenswrapper[4730]: E0221 00:06:52.434991 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.436144 4730 scope.go:117] "RemoveContainer" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" Feb 21 00:06:52 crc kubenswrapper[4730]: E0221 00:06:52.436565 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.468732 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.468796 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.468815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.468843 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.468862 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.571635 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.571709 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.571733 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.571759 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.571777 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.674711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.674756 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.674764 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.674778 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.674789 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.777479 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.777544 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.777561 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.777869 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.777948 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.880662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.880731 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.880748 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.880773 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.880790 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.984323 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.984378 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.984397 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.984420 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:52 crc kubenswrapper[4730]: I0221 00:06:52.984437 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:52Z","lastTransitionTime":"2026-02-21T00:06:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.087791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.087856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.087867 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.087890 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.087924 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.191290 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.191372 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.191390 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.191417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.191437 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.294990 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.295082 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.295101 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.295123 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.295160 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.416476 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 03:19:01.903578673 +0000 UTC Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.416870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.416964 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.416988 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.417014 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.417033 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.521386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.521456 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.521475 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.521502 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.521521 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.624381 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.624467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.624491 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.624527 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.624552 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.727756 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.727832 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.727852 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.727883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.727935 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.831582 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.831642 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.831654 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.831674 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.831689 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.934192 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.934252 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.934263 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.934287 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:53 crc kubenswrapper[4730]: I0221 00:06:53.934300 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:53Z","lastTransitionTime":"2026-02-21T00:06:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.037225 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.037270 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.037285 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.037328 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.037342 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.140615 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.140673 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.140692 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.140716 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.140732 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.243574 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.243648 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.243673 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.243700 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.243722 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.346230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.346287 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.346308 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.346333 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.346353 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.417089 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-27 15:38:35.044758098 +0000 UTC Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.434739 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.434864 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.434929 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.434991 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.435131 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.435177 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.435337 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.435605 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.448405 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.448446 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.448458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.448475 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.448487 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.552592 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.552679 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.552699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.552729 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.552749 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.655752 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.655837 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.655856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.655885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.655937 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.758980 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.759081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.759101 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.759166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.759186 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.828961 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.829094 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:54 crc kubenswrapper[4730]: E0221 00:06:54.829147 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:07:26.829130352 +0000 UTC m=+99.119966082 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.862137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.862219 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.862238 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.862266 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.862286 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.966626 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.966686 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.966706 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.966732 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:54 crc kubenswrapper[4730]: I0221 00:06:54.966750 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:54Z","lastTransitionTime":"2026-02-21T00:06:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.070358 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.070427 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.070447 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.070656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.070686 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.072291 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.072386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.072403 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.072430 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.072449 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.091470 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.098778 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.098954 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.098980 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.099047 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.099080 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.122008 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.127217 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.127292 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.127318 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.127351 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.127377 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.152396 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.157396 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.157514 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.157538 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.157565 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.157614 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.179204 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.184617 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.184704 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.184725 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.184781 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.184804 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.198534 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:55Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:55 crc kubenswrapper[4730]: E0221 00:06:55.198760 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.201061 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.201114 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.201134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.201161 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.201181 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.304731 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.304798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.304816 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.304844 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.304865 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.407930 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.407994 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.408016 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.408042 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.408062 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.417584 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-22 15:14:39.315685386 +0000 UTC Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.511461 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.511509 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.511518 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.511536 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.511546 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.619487 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.619537 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.619546 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.619562 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.619573 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.722475 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.722520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.722529 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.722548 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.722561 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.825449 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.825495 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.825505 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.825520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.825531 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.929216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.929325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.929340 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.929365 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.929379 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:55Z","lastTransitionTime":"2026-02-21T00:06:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.997653 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/0.log" Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.997713 4730 generic.go:334] "Generic (PLEG): container finished" podID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" containerID="71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd" exitCode=1 Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.997752 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerDied","Data":"71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd"} Feb 21 00:06:55 crc kubenswrapper[4730]: I0221 00:06:55.998317 4730 scope.go:117] "RemoveContainer" containerID="71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.018206 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.031373 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.031417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.031428 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.031445 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.031459 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.032496 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.044802 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.057898 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.071588 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.084972 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.101262 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.133399 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.136098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.136228 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.136325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.136416 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.136493 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.146199 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.157935 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.174312 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.185749 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.195660 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.222256 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.236877 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.239538 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.239627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.239641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.239664 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.239678 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.248395 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.259688 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"2026-02-21T00:06:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3\\\\n2026-02-21T00:06:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3 to /host/opt/cni/bin/\\\\n2026-02-21T00:06:10Z [verbose] multus-daemon started\\\\n2026-02-21T00:06:10Z [verbose] Readiness Indicator file check\\\\n2026-02-21T00:06:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:56Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.343260 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.343308 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.343325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.343342 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.343354 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.418546 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-01 18:52:41.329700108 +0000 UTC Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.434970 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.434998 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.434998 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.435108 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:56 crc kubenswrapper[4730]: E0221 00:06:56.435128 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:56 crc kubenswrapper[4730]: E0221 00:06:56.435288 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:56 crc kubenswrapper[4730]: E0221 00:06:56.435392 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:56 crc kubenswrapper[4730]: E0221 00:06:56.435505 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.445093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.445129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.445140 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.445155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.445167 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.548652 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.548700 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.548712 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.548731 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.548743 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.652394 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.652446 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.652459 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.652480 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.652494 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.756231 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.756290 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.756305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.756329 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.756344 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.858718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.858763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.858774 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.858796 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.858807 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.961471 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.961511 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.961519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.961534 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:56 crc kubenswrapper[4730]: I0221 00:06:56.961545 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:56Z","lastTransitionTime":"2026-02-21T00:06:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.004134 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/0.log" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.004246 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerStarted","Data":"ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.036293 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.060521 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.067375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.067417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.067427 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.067448 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.067460 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.077137 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.097606 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.113642 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"2026-02-21T00:06:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3\\\\n2026-02-21T00:06:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3 to /host/opt/cni/bin/\\\\n2026-02-21T00:06:10Z [verbose] multus-daemon started\\\\n2026-02-21T00:06:10Z [verbose] Readiness Indicator file check\\\\n2026-02-21T00:06:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.133194 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.153753 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.170765 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.170835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.170923 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.170953 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.170972 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.174082 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.189593 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.203664 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.218042 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.240997 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.265082 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.273897 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.273973 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.273990 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.274009 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.274024 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.284400 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.302730 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.319672 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.335286 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:57Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.377204 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.377258 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.377305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.377324 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.377338 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.420954 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 16:56:14.250556746 +0000 UTC Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.480490 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.480570 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.480597 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.480630 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.480649 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.584325 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.584389 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.584410 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.584438 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.584459 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.687482 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.687540 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.687560 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.687617 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.687636 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.791458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.791518 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.791557 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.791582 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.791598 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.895122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.895195 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.895214 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.895242 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.895264 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.998178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.998242 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.998257 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.998280 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:57 crc kubenswrapper[4730]: I0221 00:06:57.998296 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:57Z","lastTransitionTime":"2026-02-21T00:06:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.101212 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.101261 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.101274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.101290 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.101300 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.204738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.204812 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.204835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.204867 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.204889 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.308197 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.308268 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.308285 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.308311 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.308329 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.412459 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.412519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.412530 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.412551 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.412564 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.421944 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 21:13:42.735404762 +0000 UTC Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.434446 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.434490 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.434461 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.434615 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:06:58 crc kubenswrapper[4730]: E0221 00:06:58.434726 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:06:58 crc kubenswrapper[4730]: E0221 00:06:58.434834 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:06:58 crc kubenswrapper[4730]: E0221 00:06:58.435004 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:06:58 crc kubenswrapper[4730]: E0221 00:06:58.435124 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.455802 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.476623 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.489507 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.510343 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.515714 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.515749 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.515758 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.515772 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.515783 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.524678 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.536666 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.555059 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.574605 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.599308 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.614601 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.619886 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.619935 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.619945 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.619962 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.619973 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.629572 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.642101 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.667716 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.688539 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.707697 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.723007 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.723053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.723066 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.723118 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.723132 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.727163 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.745248 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"2026-02-21T00:06:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3\\\\n2026-02-21T00:06:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3 to /host/opt/cni/bin/\\\\n2026-02-21T00:06:10Z [verbose] multus-daemon started\\\\n2026-02-21T00:06:10Z [verbose] Readiness Indicator file check\\\\n2026-02-21T00:06:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:58Z is after 2025-08-24T17:21:41Z" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.825688 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.825740 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.825750 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.825769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.825779 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.928854 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.928924 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.928941 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.928961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:58 crc kubenswrapper[4730]: I0221 00:06:58.928976 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:58Z","lastTransitionTime":"2026-02-21T00:06:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.031680 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.031730 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.031740 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.031762 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.031778 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.135358 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.135445 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.135469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.135503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.135523 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.239110 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.239191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.239210 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.239236 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.239256 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.343593 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.343846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.343871 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.343961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.343991 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.422964 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-25 00:20:50.451509144 +0000 UTC Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.447700 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.447806 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.447825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.447854 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.447876 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.551481 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.551550 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.551563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.551598 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.551614 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.654353 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.654402 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.654414 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.654434 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.654450 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.757585 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.757674 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.757693 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.757721 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.757737 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.861278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.861348 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.861366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.861392 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.861413 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.964315 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.964404 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.964417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.964441 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:06:59 crc kubenswrapper[4730]: I0221 00:06:59.964457 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:06:59Z","lastTransitionTime":"2026-02-21T00:06:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.067005 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.067057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.067068 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.067086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.067096 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.169251 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.169288 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.169299 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.169314 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.169324 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.272086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.272129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.272137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.272155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.272169 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.375024 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.375126 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.375135 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.375150 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.375158 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.423157 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-19 12:44:09.838943391 +0000 UTC Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.434111 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.434164 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.434111 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:00 crc kubenswrapper[4730]: E0221 00:07:00.434251 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:00 crc kubenswrapper[4730]: E0221 00:07:00.434342 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:00 crc kubenswrapper[4730]: E0221 00:07:00.434411 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.434533 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:00 crc kubenswrapper[4730]: E0221 00:07:00.434580 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.478662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.478726 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.478752 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.478788 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.478811 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.581427 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.581462 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.581471 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.581488 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.581497 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.685885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.686015 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.686037 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.686066 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.686087 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.788835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.788870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.788880 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.788895 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.788920 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.892010 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.892069 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.892081 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.892101 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.892114 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.995464 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.995519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.995533 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.995572 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:00 crc kubenswrapper[4730]: I0221 00:07:00.995604 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:00Z","lastTransitionTime":"2026-02-21T00:07:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.099378 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.099469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.099490 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.100262 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.100327 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.203745 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.203797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.203810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.203827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.203839 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.306748 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.306812 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.306826 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.306857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.306876 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.410539 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.410593 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.410604 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.410625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.410637 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.424409 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-26 09:58:44.61199866 +0000 UTC Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.513050 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.513106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.513119 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.513139 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.513151 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.616179 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.616254 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.616264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.616286 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.616299 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.719254 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.719335 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.719345 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.719363 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.719376 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.822978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.823061 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.823091 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.823130 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.823192 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.927287 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.927342 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.927355 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.927379 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:01 crc kubenswrapper[4730]: I0221 00:07:01.927392 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:01Z","lastTransitionTime":"2026-02-21T00:07:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.030270 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.030319 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.030332 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.030352 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.030371 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.133750 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.133829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.133844 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.133870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.133887 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.236593 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.236644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.236655 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.236672 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.236683 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.340221 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.340273 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.340286 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.340305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.340319 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.425281 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-23 19:17:44.586555719 +0000 UTC Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.434844 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.435087 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:02 crc kubenswrapper[4730]: E0221 00:07:02.435184 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.435227 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.435301 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:02 crc kubenswrapper[4730]: E0221 00:07:02.435427 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:02 crc kubenswrapper[4730]: E0221 00:07:02.435535 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:02 crc kubenswrapper[4730]: E0221 00:07:02.435767 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.442609 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.442662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.442677 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.442694 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.442706 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.545275 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.545339 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.545349 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.545363 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.545375 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.649306 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.649370 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.649386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.649407 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.649428 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.752284 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.752366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.752392 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.752424 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.752448 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.855846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.855978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.855999 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.856044 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.856087 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.959968 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.960048 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.960067 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.960094 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:02 crc kubenswrapper[4730]: I0221 00:07:02.960117 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:02Z","lastTransitionTime":"2026-02-21T00:07:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.063730 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.063765 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.063775 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.063788 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.063799 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.168037 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.168153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.168176 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.168203 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.168223 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.272042 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.272114 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.272132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.272160 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.272185 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.376437 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.376548 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.376568 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.376595 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.376614 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.426309 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-16 18:55:17.887572567 +0000 UTC Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.490501 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.490578 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.490596 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.490623 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.490642 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.594587 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.594658 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.594678 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.594704 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.594727 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.697998 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.698075 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.698097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.698121 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.698144 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.801557 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.801628 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.801648 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.801679 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.801700 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.904396 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.904467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.904483 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.904511 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:03 crc kubenswrapper[4730]: I0221 00:07:03.904533 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:03Z","lastTransitionTime":"2026-02-21T00:07:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.007604 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.007645 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.007653 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.007669 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.007678 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.111025 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.111072 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.111098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.111123 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.111139 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.215546 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.215613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.215634 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.215661 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.215679 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.319765 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.319832 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.319860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.319892 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.319956 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.423176 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.423241 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.423250 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.423265 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.423276 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.427589 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 13:35:36.681588363 +0000 UTC Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.434024 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.434098 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:04 crc kubenswrapper[4730]: E0221 00:07:04.434218 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.434311 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.434043 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:04 crc kubenswrapper[4730]: E0221 00:07:04.434378 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:04 crc kubenswrapper[4730]: E0221 00:07:04.434507 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:04 crc kubenswrapper[4730]: E0221 00:07:04.434631 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.527708 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.527778 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.527791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.527815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.527829 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.631058 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.631137 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.631160 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.631189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.631211 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.734974 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.735048 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.735073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.735101 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.735123 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.838324 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.838382 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.838398 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.838421 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.838510 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.941736 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.941794 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.941807 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.941829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:04 crc kubenswrapper[4730]: I0221 00:07:04.941846 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:04Z","lastTransitionTime":"2026-02-21T00:07:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.045338 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.045412 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.045433 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.045458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.045478 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.148956 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.149053 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.149072 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.149097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.149173 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.252748 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.252835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.252859 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.252891 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.252963 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.271054 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.271126 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.271145 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.271172 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.271192 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.294762 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.301086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.301170 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.301189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.301214 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.301232 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.319699 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.325939 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.326010 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.326032 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.326063 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.326084 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.347864 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.353983 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.354042 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.354513 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.354785 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.355067 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.377375 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.383691 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.383763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.383783 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.383811 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.383829 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.403854 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:05Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:05Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:05 crc kubenswrapper[4730]: E0221 00:07:05.404165 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.406767 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.406825 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.406844 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.406872 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.406891 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.428210 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-12 11:51:21.607532519 +0000 UTC Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.510419 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.510503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.510531 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.510564 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.510591 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.613166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.613216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.613233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.613256 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.613275 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.718317 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.718396 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.718415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.718444 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.718466 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.821752 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.821810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.821828 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.821857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.821876 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.926363 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.926437 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.926456 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.926485 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:05 crc kubenswrapper[4730]: I0221 00:07:05.926512 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:05Z","lastTransitionTime":"2026-02-21T00:07:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.029824 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.029886 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.029927 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.029985 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.030005 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.132978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.133037 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.133054 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.133084 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.133105 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.237089 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.237141 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.237157 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.237181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.237201 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.341260 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.341356 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.341780 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.341861 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.342141 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.429176 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-02 08:47:14.540559379 +0000 UTC Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.434717 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.434782 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:06 crc kubenswrapper[4730]: E0221 00:07:06.434947 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.435052 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.435056 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:06 crc kubenswrapper[4730]: E0221 00:07:06.435276 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:06 crc kubenswrapper[4730]: E0221 00:07:06.435406 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:06 crc kubenswrapper[4730]: E0221 00:07:06.435560 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.445594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.445656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.445675 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.445703 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.445725 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.549069 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.549134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.549153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.549181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.549201 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.652457 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.652512 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.652525 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.652546 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.652562 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.756199 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.756271 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.756295 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.756328 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.756350 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.860099 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.860166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.860178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.860198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.860211 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.964072 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.964135 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.964147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.964166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:06 crc kubenswrapper[4730]: I0221 00:07:06.964179 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:06Z","lastTransitionTime":"2026-02-21T00:07:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.068042 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.068132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.068151 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.068179 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.068199 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.171177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.171253 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.171270 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.171295 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.171313 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.275450 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.275516 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.275538 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.275564 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.275585 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.379546 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.379622 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.379642 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.379673 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.379701 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.430067 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 08:02:58.014165052 +0000 UTC Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.436191 4730 scope.go:117] "RemoveContainer" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.484032 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.484110 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.484131 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.484164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.484188 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.587961 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.588054 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.588082 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.588115 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.588143 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.692667 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.692737 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.692753 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.692779 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.692796 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.796010 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.796046 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.796057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.796073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.796084 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.899542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.899614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.899631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.899659 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:07 crc kubenswrapper[4730]: I0221 00:07:07.899687 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:07Z","lastTransitionTime":"2026-02-21T00:07:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.002406 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.002457 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.002473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.002493 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.002508 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.056166 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/2.log" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.059425 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.059929 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.088805 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.104769 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.105535 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.105573 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.105584 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.105602 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.105615 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.120708 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.132118 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.143238 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.160533 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.184616 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.199925 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.208282 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.208352 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.208374 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.208403 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.208421 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.223705 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.238979 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.254394 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"2026-02-21T00:06:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3\\\\n2026-02-21T00:06:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3 to /host/opt/cni/bin/\\\\n2026-02-21T00:06:10Z [verbose] multus-daemon started\\\\n2026-02-21T00:06:10Z [verbose] Readiness Indicator file check\\\\n2026-02-21T00:06:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.272151 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.285502 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.302870 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.311937 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.311983 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.311993 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.312008 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.312019 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.322929 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.346407 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.358643 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.415159 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.415216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.415230 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.415251 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.415266 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.430808 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-17 13:19:00.541206744 +0000 UTC Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.436323 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.436369 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.436327 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.436315 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:08 crc kubenswrapper[4730]: E0221 00:07:08.436466 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:08 crc kubenswrapper[4730]: E0221 00:07:08.436731 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:08 crc kubenswrapper[4730]: E0221 00:07:08.436959 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:08 crc kubenswrapper[4730]: E0221 00:07:08.437153 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.452627 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://92bfbf0a96c581d73da9f4d13adeeb4462754cbdaf52115ae6e71a0fe34ef3fa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.464853 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c60ab99969ba83c2410ab868f4b310b6c2b37598621c31edea4b5664f7c7b92c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://6cfcb592fdd9cb9f7b6f93f8b2213877a627e5cc81d07a1cfcd167ffd46db0f5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.476519 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"33513de3-5480-4aef-87ff-879f9e7a475f\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8ba365963cb762a286e6e08bc1264db75dcbdc5cd2e8c5660aa9a39b73361244\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9cl57\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-r8q7m\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.485618 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-krrw8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rjwnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-krrw8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.506958 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"936a78db-aaaa-4052-9d67-e85e4dbd03a1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb3e3113c566024c0efc1315b0c83ef5a1064f961d024f0b249015e5e5fde721\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://10a16482028091d6922d49ba3c21957d0d2e44773062878e6ebbd86c09471520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d033560815d1e6134b5b432053817da793a9fa6a9ef1f9c2b1b659ed1246d770\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.518626 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.518684 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.518698 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.518718 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.518734 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.525235 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"6b7bc1c6-7005-4063-85e2-07b33c1c882c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:39Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://22e8630b37efe8908924708cd1ee669813533fb7c968b8e7e2fdaf9ab5cd3355\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c5d78d9c404dbeefc99ff272cc86c7127fbb3a2b7b6a552e721aecdd03215b12\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b1a79c29fb257c0aa7b8efdc0ba24581f7b4e03a7f324afe718e75928574daae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5917c60762939abfba1b3fe3e32b4555722a0b45b573b942777026ce52a9ec9b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.542248 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f8a36a70-028a-4dca-87d5-b4d1a8b1d048\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://095c578339c9d7bc3a38527d8a7f1059ed104a132cbdd3ed164ae001263d0ef9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ec4a163ebfc0696a903196609f42202ddc24ccb46cf6215ad8d8b9da99d1895\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:22Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tgj7l\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-k28mx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.572736 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfb5b15d-d281-4633-9930-905d05ed76e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:37Z\\\",\\\"message\\\":\\\" not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:06:37Z is after 2025-08-24T17:21:41Z]\\\\nI0221 00:06:37.476177 6372 transact.go:42] Configuring OVN: [{Op:update Table:Load_Balancer Row:map[external_ids:{GoMap:map[k8s.ovn.org/kind:Service k8s.ovn.org/owner:openshift-machine-api/cluster-autoscaler-operator]} name:Service_openshift-machine-api/cluster-autoscaler-operator_TCP_cluster options:{GoMap:map[event:false hairpin_snat_ip:169.254.0.5 fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.245:443: 10.217.5.245:9192:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {54fbe873-7e6d-475f-a0ad-8dd5f06d850d}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0221 00:06:37.476241 \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:07:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9jdnh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-rjptr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.592961 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-srs5q" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"638e980f-5038-4a5d-85d2-4aae577964af\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://571ded1f1b0a85f9de885613b0a98b7d8df46a5e952f6b1c3a1b2bdf978242f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9786b7acbaac7d310126d8e6aaa1d77231eed4808a34910cfb5091ea29f149e8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://34abef87cae56b8395365d09f00b76abbc131657b6d1902bd8ab8dd79083c883\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://61048d88445750216e87f654a97bb010d6924f6facae914989ff0bbf6fed7e76\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3847b87c09517087d9d984f16fcc5316d544c061312798bfaaafa2b3d380fc3c\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:12Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c0eff26436b3b412f800cd73feba4a9b4643c8d664176c0faaf7bc558cf55229\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c202e15591fd3a2bd6f4afff5b0760b94e047cc8ea8bb2f6e6790fd42d298ea7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:06:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6z762\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-srs5q\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.606166 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621374 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621418 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621431 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621451 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621466 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.621557 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m9nnd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ac51a893-4a0c-44ed-8284-0aac9e8d02ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:57Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-02-21T00:06:55Z\\\",\\\"message\\\":\\\"2026-02-21T00:06:10+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3\\\\n2026-02-21T00:06:10+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_9887d52a-b598-479c-8d36-7ed46998fbb3 to /host/opt/cni/bin/\\\\n2026-02-21T00:06:10Z [verbose] multus-daemon started\\\\n2026-02-21T00:06:10Z [verbose] Readiness Indicator file check\\\\n2026-02-21T00:06:55Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:06:08Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:56Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-44z99\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m9nnd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.636568 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.649644 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-wkkq5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d5fbb14d-65bb-4a3a-afc7-7c4404f0fe14\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://edf9e2ccf78693107c9bf3f131b8f8d9fd36bd61a7bc5eada134434e8e0cffbb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-lmvms\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:08Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-wkkq5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.665195 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-f7kbj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a502eb4e-6f7f-4567-a7fd-bc15eb806845\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4cda804985fb46821fdf7fde6dc379e6f3c1c3e4669ae75ed6f207fd18e84cbe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-78bbl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:06:09Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-f7kbj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.684895 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:27Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-02-21T00:05:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-02-21T00:06:07Z\\\",\\\"message\\\":\\\"-03-23 00:05:52 +0000 UTC (now=2026-02-21 00:06:07.580033122 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580178 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI0221 00:06:07.580190 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI0221 00:06:07.580197 1 named_certificates.go:53] \\\\\\\"Loaded SNI cert\\\\\\\" index=0 certName=\\\\\\\"self-signed loopback\\\\\\\" certDetail=\\\\\\\"\\\\\\\\\\\\\\\"apiserver-loopback-client@1771632362\\\\\\\\\\\\\\\" [serving] validServingFor=[apiserver-loopback-client] issuer=\\\\\\\\\\\\\\\"apiserver-loopback-client-ca@1771632362\\\\\\\\\\\\\\\" (2026-02-20 23:06:02 +0000 UTC to 2027-02-20 23:06:02 +0000 UTC (now=2026-02-21 00:06:07.580172086 +0000 UTC))\\\\\\\"\\\\nI0221 00:06:07.580217 1 secure_serving.go:213] Serving securely on [::]:17697\\\\nI0221 00:06:07.580245 1 genericapiserver.go:683] [graceful-termination] waiting for shutdown to be initiated\\\\nI0221 00:06:07.580269 1 dynamic_serving_content.go:135] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-1258452935/tls.crt::/tmp/serving-cert-1258452935/tls.key\\\\\\\"\\\\nI0221 00:06:07.580079 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI0221 00:06:07.580398 1 tlsconfig.go:243] \\\\\\\"Starting DynamicServingCertificateController\\\\\\\"\\\\nI0221 00:06:07.580420 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580432 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI0221 00:06:07.580450 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nI0221 00:06:07.580457 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\nF0221 00:06:07.580590 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:05:51Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-02-21T00:05:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-02-21T00:05:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-02-21T00:05:48Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.701384 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:09Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f99baa3a8c3400ea8386699f307e6a3ccd653ff8f053793589e38c6e5fd38d7c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-02-21T00:06:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.713029 4730 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-02-21T00:06:08Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:08Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.724239 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.724303 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.724327 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.724363 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.724385 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.827642 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.827688 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.827697 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.827713 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.827729 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.931223 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.931337 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.931362 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.931386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:08 crc kubenswrapper[4730]: I0221 00:07:08.931406 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:08Z","lastTransitionTime":"2026-02-21T00:07:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.034067 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.034113 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.034122 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.034138 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.034152 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.137406 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.137848 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.137867 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.137894 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.137941 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.248792 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.248883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.248963 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.248993 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.249012 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.352365 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.352441 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.352463 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.352495 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.353008 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.431005 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-12 12:03:25.794563573 +0000 UTC Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.457073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.457146 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.457166 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.457194 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.457215 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.560779 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.560846 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.560865 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.560893 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.560950 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.664938 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.665000 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.665019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.665049 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.665070 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.768624 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.768683 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.768702 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.768730 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.768751 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.872035 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.872087 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.872105 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.872129 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.872148 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.976226 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.976268 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.976278 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.976295 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:09 crc kubenswrapper[4730]: I0221 00:07:09.976308 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:09Z","lastTransitionTime":"2026-02-21T00:07:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.078848 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.078947 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.078973 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.079005 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.079030 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.182757 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.182817 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.182837 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.182861 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.182880 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.286798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.286873 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.286893 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.286970 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.286994 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.390405 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.391004 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.391189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.391339 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.391481 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.431452 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-28 08:16:24.82020118 +0000 UTC Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.434984 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.435213 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.435117 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.435117 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:10 crc kubenswrapper[4730]: E0221 00:07:10.435635 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:10 crc kubenswrapper[4730]: E0221 00:07:10.435813 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:10 crc kubenswrapper[4730]: E0221 00:07:10.436015 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:10 crc kubenswrapper[4730]: E0221 00:07:10.436155 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.494448 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.494798 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.495105 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.495308 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.495530 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.599181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.599557 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.599698 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.599850 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.600040 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.703775 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.704480 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.704699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.704885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.705089 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.808790 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.809227 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.809361 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.809540 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.809678 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.913374 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.913437 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.913455 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.913479 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:10 crc kubenswrapper[4730]: I0221 00:07:10.913497 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:10Z","lastTransitionTime":"2026-02-21T00:07:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.016845 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.016936 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.016959 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.016993 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.017016 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.120250 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.120309 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.120330 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.120354 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.120378 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.223810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.223864 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.223885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.223945 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.223969 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.327621 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.327685 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.327703 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.327728 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.327746 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.431222 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.431668 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.432453 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.431747 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-15 02:19:30.781021887 +0000 UTC Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.432605 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.432679 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.536999 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.537168 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.537198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.537233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.537260 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.640413 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.640484 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.640502 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.640532 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.640550 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.744098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.744178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.744204 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.744237 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.744263 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.847359 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.847430 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.847449 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.847476 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.847500 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.950834 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.950925 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.950945 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.950973 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:11 crc kubenswrapper[4730]: I0221 00:07:11.950992 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:11Z","lastTransitionTime":"2026-02-21T00:07:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.055063 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.055140 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.055157 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.055365 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.055391 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.158655 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.158734 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.158753 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.158788 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.158809 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.263239 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.263331 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.263365 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.263415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.263441 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.340273 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.340410 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340643 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340752 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340777 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340869 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:16.340838798 +0000 UTC m=+148.631674558 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340655 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340968 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.340985 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.341033 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:16.341019733 +0000 UTC m=+148.631855493 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.367553 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.367628 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.367654 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.367688 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.367713 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.433443 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-14 05:31:31.870862397 +0000 UTC Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.435250 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.435341 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.435505 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.435707 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.435684 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.435832 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.435989 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.436116 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.441304 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.441517 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:16.441486931 +0000 UTC m=+148.732322671 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.470270 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.470326 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.470345 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.470375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.470395 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.542675 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.542795 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.543037 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.543136 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:16.543110664 +0000 UTC m=+148.833946424 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.543181 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: E0221 00:07:12.543478 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:16.543434353 +0000 UTC m=+148.834270253 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.575242 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.575346 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.575372 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.575439 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.575465 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.679078 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.679165 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.679202 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.679232 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.679255 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.782272 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.782324 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.782341 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.782366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.782384 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.886189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.886255 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.886269 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.886292 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.886310 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.990432 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.990517 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.990535 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.990563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:12 crc kubenswrapper[4730]: I0221 00:07:12.990585 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:12Z","lastTransitionTime":"2026-02-21T00:07:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.094659 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.094716 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.094737 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.094763 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.094781 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.199018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.199123 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.199153 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.199187 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.199215 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.302722 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.302817 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.302839 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.302872 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.302895 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.406424 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.406502 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.406520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.406549 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.406568 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.434307 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-19 17:34:13.198247359 +0000 UTC Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.510560 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.510625 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.510647 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.510674 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.510694 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.614155 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.614257 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.614274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.614299 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.614321 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.717872 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.717979 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.717998 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.718024 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.718068 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.821991 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.822152 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.822187 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.822257 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.822284 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.925997 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.926092 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.926143 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.926170 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:13 crc kubenswrapper[4730]: I0221 00:07:13.926188 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:13Z","lastTransitionTime":"2026-02-21T00:07:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.031134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.031224 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.031246 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.031276 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.031300 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.134554 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.134649 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.134675 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.134715 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.134743 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.238641 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.238720 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.238743 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.238772 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.238792 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.342461 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.342542 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.342577 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.342616 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.342640 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.434290 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.434336 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.434407 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.434323 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.434485 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-11 13:33:47.239928596 +0000 UTC Feb 21 00:07:14 crc kubenswrapper[4730]: E0221 00:07:14.434548 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:14 crc kubenswrapper[4730]: E0221 00:07:14.434690 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:14 crc kubenswrapper[4730]: E0221 00:07:14.434880 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:14 crc kubenswrapper[4730]: E0221 00:07:14.435027 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.445185 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.445269 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.445344 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.445375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.445397 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.548636 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.548715 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.548733 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.548759 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.548778 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.653563 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.653636 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.653661 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.653692 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.653715 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.758019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.758088 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.758106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.758132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.758155 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.861092 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.861161 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.861178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.861237 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.861256 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.965100 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.965178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.965201 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.965237 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:14 crc kubenswrapper[4730]: I0221 00:07:14.965260 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:14Z","lastTransitionTime":"2026-02-21T00:07:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.068434 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.068885 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.068947 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.068975 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.068993 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.172223 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.172383 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.172416 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.172447 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.172475 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.275764 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.275838 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.275858 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.275888 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.275936 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.379078 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.379132 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.379148 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.379172 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.379191 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.434699 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-08 14:31:36.146735126 +0000 UTC Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.483613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.483692 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.483713 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.483745 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.483771 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.587069 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.587165 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.587189 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.587219 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.587240 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.597884 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.598093 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.598113 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.598135 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.598153 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.619892 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.625014 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.625086 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.625106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.625134 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.625154 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.645604 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.652085 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.652147 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.652167 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.652199 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.652219 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.673548 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.681605 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.681708 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.681769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.681802 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.681862 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.704645 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.710414 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.710473 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.710494 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.710521 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.710540 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.730641 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"lastTransitionTime\\\":\\\"2026-02-21T00:07:15Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"38c4a1be-5e86-41d3-8a2e-4b0c3d994a3b\\\",\\\"systemUUID\\\":\\\"2c908926-b11f-4f79-baa4-c8fb64a7edb3\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-02-21T00:07:15Z is after 2025-08-24T17:21:41Z" Feb 21 00:07:15 crc kubenswrapper[4730]: E0221 00:07:15.731037 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.733491 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.733574 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.733595 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.733630 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.733656 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.836729 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.836811 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.836831 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.836863 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.836890 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.940767 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.940835 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.940852 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.940878 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:15 crc kubenswrapper[4730]: I0221 00:07:15.940896 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:15Z","lastTransitionTime":"2026-02-21T00:07:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.044262 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.044328 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.044346 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.044371 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.044390 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.147519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.147620 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.147652 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.147689 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.147715 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.251269 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.251340 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.251363 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.251395 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.251420 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.355106 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.355164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.355181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.355208 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.355227 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.435067 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-09 02:56:58.251598445 +0000 UTC Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.435164 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.435217 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.435259 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.435379 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:16 crc kubenswrapper[4730]: E0221 00:07:16.435967 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:16 crc kubenswrapper[4730]: E0221 00:07:16.436049 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:16 crc kubenswrapper[4730]: E0221 00:07:16.436125 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:16 crc kubenswrapper[4730]: E0221 00:07:16.436262 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.458159 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.458216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.458234 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.458258 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.458280 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.561128 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.561164 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.561175 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.561191 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.561208 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.663633 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.663711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.663741 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.663772 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.663795 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.767184 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.767329 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.767352 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.767382 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.767450 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.870522 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.870629 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.870656 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.870689 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.870716 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.974104 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.974177 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.974199 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.974229 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:16 crc kubenswrapper[4730]: I0221 00:07:16.974250 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:16Z","lastTransitionTime":"2026-02-21T00:07:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.076514 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.076592 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.076612 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.076640 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.076661 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.179805 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.179868 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.179886 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.179936 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.179955 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.283743 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.283797 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.283807 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.283826 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.283840 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.387199 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.387289 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.387304 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.387331 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.387347 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.435562 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-06 11:19:26.409035168 +0000 UTC Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.490815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.490962 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.490986 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.491018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.491037 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.594836 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.594934 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.594953 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.594980 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.595000 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.698864 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.698960 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.698985 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.699019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.699044 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.803118 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.803224 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.803255 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.803293 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.803320 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.907528 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.907583 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.907600 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.907629 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:17 crc kubenswrapper[4730]: I0221 00:07:17.907653 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:17Z","lastTransitionTime":"2026-02-21T00:07:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.011462 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.011543 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.011562 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.011588 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.011607 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.115098 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.115168 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.115190 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.115219 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.115239 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.218527 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.218658 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.218685 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.218711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.218767 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.321628 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.321696 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.321713 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.321739 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.321767 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.425431 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.425485 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.425504 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.425530 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.425548 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.434139 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.434195 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:18 crc kubenswrapper[4730]: E0221 00:07:18.434287 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.434330 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.434443 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:18 crc kubenswrapper[4730]: E0221 00:07:18.434661 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:18 crc kubenswrapper[4730]: E0221 00:07:18.434728 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:18 crc kubenswrapper[4730]: E0221 00:07:18.434932 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.436485 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-22 19:49:20.692561398 +0000 UTC Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.448829 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.525872 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podStartSLOduration=71.525839438 podStartE2EDuration="1m11.525839438s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.491773685 +0000 UTC m=+90.782609455" watchObservedRunningTime="2026-02-21 00:07:18.525839438 +0000 UTC m=+90.816675198" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.530552 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.530652 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.530721 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.530759 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.530839 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.574854 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-srs5q" podStartSLOduration=71.57478971 podStartE2EDuration="1m11.57478971s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.526231899 +0000 UTC m=+90.817067699" watchObservedRunningTime="2026-02-21 00:07:18.57478971 +0000 UTC m=+90.865625480" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.579837 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-k28mx" podStartSLOduration=70.579757773 podStartE2EDuration="1m10.579757773s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.573556705 +0000 UTC m=+90.864392455" watchObservedRunningTime="2026-02-21 00:07:18.579757773 +0000 UTC m=+90.870593533" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.628534 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-m9nnd" podStartSLOduration=71.6285048 podStartE2EDuration="1m11.6285048s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.606934278 +0000 UTC m=+90.897770018" watchObservedRunningTime="2026-02-21 00:07:18.6285048 +0000 UTC m=+90.919340550" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.635107 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.635175 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.635199 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.635228 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.635246 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.730655 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-f7kbj" podStartSLOduration=71.730609066 podStartE2EDuration="1m11.730609066s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.730111832 +0000 UTC m=+91.020947612" watchObservedRunningTime="2026-02-21 00:07:18.730609066 +0000 UTC m=+91.021444846" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.731465 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-wkkq5" podStartSLOduration=71.73144723 podStartE2EDuration="1m11.73144723s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.717183599 +0000 UTC m=+91.008019369" watchObservedRunningTime="2026-02-21 00:07:18.73144723 +0000 UTC m=+91.022283000" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.739523 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.739579 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.739594 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.739617 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.739633 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.795674 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=70.795641013 podStartE2EDuration="1m10.795641013s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.764831773 +0000 UTC m=+91.055667523" watchObservedRunningTime="2026-02-21 00:07:18.795641013 +0000 UTC m=+91.086476763" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.825630 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=64.825589217 podStartE2EDuration="1m4.825589217s" podCreationTimestamp="2026-02-21 00:06:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.800573895 +0000 UTC m=+91.091409675" watchObservedRunningTime="2026-02-21 00:07:18.825589217 +0000 UTC m=+91.116424947" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.826092 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=39.826087641 podStartE2EDuration="39.826087641s" podCreationTimestamp="2026-02-21 00:06:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.825269637 +0000 UTC m=+91.116105367" watchObservedRunningTime="2026-02-21 00:07:18.826087641 +0000 UTC m=+91.116923371" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.842232 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.842269 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.842279 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.842293 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.842307 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.869546 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podStartSLOduration=71.869527104 podStartE2EDuration="1m11.869527104s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:18.869135333 +0000 UTC m=+91.159971063" watchObservedRunningTime="2026-02-21 00:07:18.869527104 +0000 UTC m=+91.160362834" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.946644 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.946739 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.946758 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.946786 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:18 crc kubenswrapper[4730]: I0221 00:07:18.946834 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:18Z","lastTransitionTime":"2026-02-21T00:07:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.050786 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.050860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.050877 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.050931 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.050989 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.154087 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.154210 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.154233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.154268 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.154288 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.258023 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.258094 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.258111 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.258141 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.258159 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.362121 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.362181 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.362198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.362223 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.362242 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.437169 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-20 07:00:20.918221028 +0000 UTC Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.465827 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.465890 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.465948 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.465978 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.465996 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.570019 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.570083 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.570099 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.570124 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.570143 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.674180 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.674281 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.674312 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.674346 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.674370 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.778162 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.778242 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.778264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.778291 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.778309 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.882070 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.882178 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.882200 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.882227 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.882245 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.985524 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.985589 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.985613 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.985647 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:19 crc kubenswrapper[4730]: I0221 00:07:19.985670 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:19Z","lastTransitionTime":"2026-02-21T00:07:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.088190 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.088253 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.088271 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.088298 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.088317 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.190810 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.190889 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.190941 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.190965 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.191000 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.295219 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.295298 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.295318 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.295344 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.295364 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.399057 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.399138 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.399162 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.399193 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.399216 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.434336 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.434381 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.434390 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:20 crc kubenswrapper[4730]: E0221 00:07:20.434560 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.434588 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:20 crc kubenswrapper[4730]: E0221 00:07:20.434757 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:20 crc kubenswrapper[4730]: E0221 00:07:20.434888 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:20 crc kubenswrapper[4730]: E0221 00:07:20.435034 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.437694 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-05 20:08:34.755839699 +0000 UTC Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.503666 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.504647 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.505052 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.505381 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.505721 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.609796 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.609887 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.609947 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.609986 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.610014 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.714026 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.714097 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.714116 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.714145 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.714166 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.817305 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.817368 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.817388 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.817415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.817436 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.920659 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.920723 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.920742 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.920769 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:20 crc kubenswrapper[4730]: I0221 00:07:20.920792 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:20Z","lastTransitionTime":"2026-02-21T00:07:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.024334 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.024395 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.024415 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.024440 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.024462 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.127276 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.127417 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.127439 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.127466 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.127483 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.230855 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.231018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.231047 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.231077 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.231102 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.334386 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.334461 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.334484 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.334514 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.334534 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.437820 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-10 08:23:46.20766515 +0000 UTC Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.438754 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.438814 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.438833 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.438858 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.438877 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.460950 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.542113 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.542198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.542217 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.542245 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.542269 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.645559 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.645649 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.645667 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.645699 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.645725 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.750265 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.750316 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.750332 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.750358 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.750380 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.853959 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.854017 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.854033 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.854058 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.854077 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.957705 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.957802 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.957829 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.957862 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:21 crc kubenswrapper[4730]: I0221 00:07:21.957886 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:21Z","lastTransitionTime":"2026-02-21T00:07:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.062467 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.062533 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.062550 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.062580 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.062603 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.166313 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.166389 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.166414 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.166449 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.166473 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.269519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.269585 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.269607 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.269640 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.269666 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.373851 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.374002 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.374039 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.374071 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.374090 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.434588 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.434712 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.434662 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.434633 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:22 crc kubenswrapper[4730]: E0221 00:07:22.435051 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:22 crc kubenswrapper[4730]: E0221 00:07:22.435213 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:22 crc kubenswrapper[4730]: E0221 00:07:22.435290 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:22 crc kubenswrapper[4730]: E0221 00:07:22.435454 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.438789 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-10 07:02:35.024705168 +0000 UTC Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.476633 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.476692 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.476711 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.476738 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.476767 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.580292 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.580379 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.580402 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.580439 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.580466 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.683747 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.683861 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.683876 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.683893 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.683926 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.787544 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.787627 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.787651 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.787703 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.787728 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.891757 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.891837 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.891857 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.891883 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.891936 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.994637 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.994708 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.994727 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.994753 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:22 crc kubenswrapper[4730]: I0221 00:07:22.994770 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:22Z","lastTransitionTime":"2026-02-21T00:07:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.098434 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.098507 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.098519 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.098543 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.098561 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.202412 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.202499 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.202525 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.202561 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.202585 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.306523 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.306608 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.306631 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.306662 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.306691 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.410781 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.410856 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.410893 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.410979 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.411001 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.439394 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-09 05:06:11.779796957 +0000 UTC Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.440664 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" probeResult="failure" output="" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.514403 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.514498 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.514522 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.514553 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.514574 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.618152 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.618229 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.618256 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.618288 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.618314 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.722349 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.722444 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.722468 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.722503 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.722544 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.826013 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.826073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.826084 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.826102 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.826113 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.929246 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.929329 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.929343 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.929366 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:23 crc kubenswrapper[4730]: I0221 00:07:23.929795 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:23Z","lastTransitionTime":"2026-02-21T00:07:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.033426 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.033506 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.033525 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.033553 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.033574 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.136994 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.137070 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.137088 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.137113 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.137134 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.243512 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.243589 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.243608 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.243638 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.243657 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.348786 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.348895 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.348963 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.348996 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.349017 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.435094 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.435183 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.435107 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.435426 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:24 crc kubenswrapper[4730]: E0221 00:07:24.435526 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:24 crc kubenswrapper[4730]: E0221 00:07:24.435675 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:24 crc kubenswrapper[4730]: E0221 00:07:24.435853 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:24 crc kubenswrapper[4730]: E0221 00:07:24.436019 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.439636 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-17 01:27:32.787695642 +0000 UTC Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.451231 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.451264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.451274 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.451290 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.451302 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.554870 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.555018 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.555043 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.555080 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.555107 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.658264 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.658598 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.658736 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.658815 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.658930 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.762381 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.762494 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.762520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.762551 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.762570 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.866571 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.867064 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.867233 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.867419 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.867611 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.971136 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.971198 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.971219 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.971246 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:24 crc kubenswrapper[4730]: I0221 00:07:24.971263 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:24Z","lastTransitionTime":"2026-02-21T00:07:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.075156 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.075216 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.075232 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.075259 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.075279 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.178866 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.179004 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.179036 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.179073 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.179100 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.283375 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.283458 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.283480 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.283508 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.283529 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.387117 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.387195 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.387217 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.387242 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.387260 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.440595 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-03 23:07:20.601033463 +0000 UTC Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.490737 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.490819 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.490843 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.490878 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.490896 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.594067 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.594131 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.594148 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.594172 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.594190 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.697791 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.697860 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.697888 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.697951 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.697980 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.804469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.804534 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.804549 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.804576 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.804599 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.907520 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.907589 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.907614 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.907648 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:25 crc kubenswrapper[4730]: I0221 00:07:25.907674 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:25Z","lastTransitionTime":"2026-02-21T00:07:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.011830 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.011934 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.011951 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.011975 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.011987 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:26Z","lastTransitionTime":"2026-02-21T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.080601 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.080672 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.080692 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.080720 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.080738 4730 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-02-21T00:07:26Z","lastTransitionTime":"2026-02-21T00:07:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.154206 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t"] Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.154854 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.158719 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.159127 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.160191 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.160455 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.178440 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=8.178408452 podStartE2EDuration="8.178408452s" podCreationTimestamp="2026-02-21 00:07:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:26.177390493 +0000 UTC m=+98.468226263" watchObservedRunningTime="2026-02-21 00:07:26.178408452 +0000 UTC m=+98.469244222" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.245748 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.245718674 podStartE2EDuration="5.245718674s" podCreationTimestamp="2026-02-21 00:07:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:26.24455512 +0000 UTC m=+98.535390930" watchObservedRunningTime="2026-02-21 00:07:26.245718674 +0000 UTC m=+98.536554444" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.319424 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.319501 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f76dc95-9da9-4968-b834-782ec3d6c05f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.319565 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.319607 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f76dc95-9da9-4968-b834-782ec3d6c05f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.320101 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4f76dc95-9da9-4968-b834-782ec3d6c05f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421537 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4f76dc95-9da9-4968-b834-782ec3d6c05f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421625 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421661 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f76dc95-9da9-4968-b834-782ec3d6c05f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421718 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421758 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f76dc95-9da9-4968-b834-782ec3d6c05f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421849 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.421985 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/4f76dc95-9da9-4968-b834-782ec3d6c05f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.423524 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4f76dc95-9da9-4968-b834-782ec3d6c05f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.434426 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.434506 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.434608 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.434627 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.435035 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.435202 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4f76dc95-9da9-4968-b834-782ec3d6c05f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.435194 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.435369 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.435520 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.440735 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-07 17:08:39.22929463 +0000 UTC Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.440813 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.453055 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4f76dc95-9da9-4968-b834-782ec3d6c05f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-2mn4t\" (UID: \"4f76dc95-9da9-4968-b834-782ec3d6c05f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.453736 4730 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.482137 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" Feb 21 00:07:26 crc kubenswrapper[4730]: W0221 00:07:26.511416 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4f76dc95_9da9_4968_b834_782ec3d6c05f.slice/crio-911e13b3f15ad59d253e84fc9ada4ade646325e47a40c1f65bce345f558646ec WatchSource:0}: Error finding container 911e13b3f15ad59d253e84fc9ada4ade646325e47a40c1f65bce345f558646ec: Status 404 returned error can't find the container with id 911e13b3f15ad59d253e84fc9ada4ade646325e47a40c1f65bce345f558646ec Feb 21 00:07:26 crc kubenswrapper[4730]: I0221 00:07:26.929200 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.929390 4730 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:07:26 crc kubenswrapper[4730]: E0221 00:07:26.929481 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs podName:73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.929455263 +0000 UTC m=+163.220291003 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs") pod "network-metrics-daemon-krrw8" (UID: "73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971") : object "openshift-multus"/"metrics-daemon-secret" not registered Feb 21 00:07:27 crc kubenswrapper[4730]: I0221 00:07:27.148180 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" event={"ID":"4f76dc95-9da9-4968-b834-782ec3d6c05f","Type":"ContainerStarted","Data":"020074cb5465908fb37f156efe455458845d7680c517344fe3f4e60ad8a0d662"} Feb 21 00:07:27 crc kubenswrapper[4730]: I0221 00:07:27.148288 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" event={"ID":"4f76dc95-9da9-4968-b834-782ec3d6c05f","Type":"ContainerStarted","Data":"911e13b3f15ad59d253e84fc9ada4ade646325e47a40c1f65bce345f558646ec"} Feb 21 00:07:27 crc kubenswrapper[4730]: I0221 00:07:27.173573 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-2mn4t" podStartSLOduration=80.173542445 podStartE2EDuration="1m20.173542445s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:07:27.171684342 +0000 UTC m=+99.462520102" watchObservedRunningTime="2026-02-21 00:07:27.173542445 +0000 UTC m=+99.464378215" Feb 21 00:07:28 crc kubenswrapper[4730]: I0221 00:07:28.436609 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:28 crc kubenswrapper[4730]: I0221 00:07:28.436954 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:28 crc kubenswrapper[4730]: E0221 00:07:28.437567 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:28 crc kubenswrapper[4730]: I0221 00:07:28.437132 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:28 crc kubenswrapper[4730]: I0221 00:07:28.437004 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:28 crc kubenswrapper[4730]: E0221 00:07:28.437731 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:28 crc kubenswrapper[4730]: E0221 00:07:28.438093 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:28 crc kubenswrapper[4730]: E0221 00:07:28.438197 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.159553 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/3.log" Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.161301 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/2.log" Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.172582 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" exitCode=1 Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.172663 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82"} Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.172730 4730 scope.go:117] "RemoveContainer" containerID="c96be72b889a3de025f4d7128c442152e04ba6b4e8558e1c82f836133778fd01" Feb 21 00:07:29 crc kubenswrapper[4730]: I0221 00:07:29.174589 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:07:29 crc kubenswrapper[4730]: E0221 00:07:29.175167 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:07:30 crc kubenswrapper[4730]: I0221 00:07:30.180792 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/3.log" Feb 21 00:07:30 crc kubenswrapper[4730]: I0221 00:07:30.437377 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:30 crc kubenswrapper[4730]: I0221 00:07:30.437397 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:30 crc kubenswrapper[4730]: I0221 00:07:30.437460 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:30 crc kubenswrapper[4730]: I0221 00:07:30.437743 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:30 crc kubenswrapper[4730]: E0221 00:07:30.437814 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:30 crc kubenswrapper[4730]: E0221 00:07:30.437922 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:30 crc kubenswrapper[4730]: E0221 00:07:30.437671 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:30 crc kubenswrapper[4730]: E0221 00:07:30.438071 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:32 crc kubenswrapper[4730]: I0221 00:07:32.434831 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:32 crc kubenswrapper[4730]: I0221 00:07:32.434942 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:32 crc kubenswrapper[4730]: I0221 00:07:32.435010 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:32 crc kubenswrapper[4730]: E0221 00:07:32.436391 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:32 crc kubenswrapper[4730]: I0221 00:07:32.435118 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:32 crc kubenswrapper[4730]: E0221 00:07:32.436604 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:32 crc kubenswrapper[4730]: E0221 00:07:32.436759 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:32 crc kubenswrapper[4730]: E0221 00:07:32.437022 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:34 crc kubenswrapper[4730]: I0221 00:07:34.434995 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:34 crc kubenswrapper[4730]: I0221 00:07:34.435051 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:34 crc kubenswrapper[4730]: I0221 00:07:34.435061 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:34 crc kubenswrapper[4730]: I0221 00:07:34.435150 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:34 crc kubenswrapper[4730]: E0221 00:07:34.435311 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:34 crc kubenswrapper[4730]: E0221 00:07:34.435720 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:34 crc kubenswrapper[4730]: E0221 00:07:34.435818 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:34 crc kubenswrapper[4730]: E0221 00:07:34.435988 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:36 crc kubenswrapper[4730]: I0221 00:07:36.434143 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:36 crc kubenswrapper[4730]: I0221 00:07:36.434248 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:36 crc kubenswrapper[4730]: I0221 00:07:36.434404 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:36 crc kubenswrapper[4730]: E0221 00:07:36.434629 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:36 crc kubenswrapper[4730]: I0221 00:07:36.434724 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:36 crc kubenswrapper[4730]: E0221 00:07:36.434806 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:36 crc kubenswrapper[4730]: E0221 00:07:36.434985 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:36 crc kubenswrapper[4730]: E0221 00:07:36.435105 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:38 crc kubenswrapper[4730]: I0221 00:07:38.434761 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:38 crc kubenswrapper[4730]: I0221 00:07:38.434834 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:38 crc kubenswrapper[4730]: I0221 00:07:38.434890 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:38 crc kubenswrapper[4730]: I0221 00:07:38.437327 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:38 crc kubenswrapper[4730]: E0221 00:07:38.437488 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:38 crc kubenswrapper[4730]: E0221 00:07:38.437629 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:38 crc kubenswrapper[4730]: E0221 00:07:38.437793 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:38 crc kubenswrapper[4730]: E0221 00:07:38.437955 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:40 crc kubenswrapper[4730]: I0221 00:07:40.434833 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:40 crc kubenswrapper[4730]: I0221 00:07:40.434962 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:40 crc kubenswrapper[4730]: I0221 00:07:40.435033 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:40 crc kubenswrapper[4730]: E0221 00:07:40.435325 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:40 crc kubenswrapper[4730]: I0221 00:07:40.435367 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:40 crc kubenswrapper[4730]: E0221 00:07:40.435505 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:40 crc kubenswrapper[4730]: E0221 00:07:40.435657 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:40 crc kubenswrapper[4730]: E0221 00:07:40.435766 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:41 crc kubenswrapper[4730]: I0221 00:07:41.435708 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:07:41 crc kubenswrapper[4730]: E0221 00:07:41.436036 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.237332 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/1.log" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.238111 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/0.log" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.238167 4730 generic.go:334] "Generic (PLEG): container finished" podID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" containerID="ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6" exitCode=1 Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.238210 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerDied","Data":"ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6"} Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.238256 4730 scope.go:117] "RemoveContainer" containerID="71bfaa52eb65994d27419d507e4364a9f6c35f4ab7b3e6567234f33bee2c15cd" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.238821 4730 scope.go:117] "RemoveContainer" containerID="ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6" Feb 21 00:07:42 crc kubenswrapper[4730]: E0221 00:07:42.239107 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-m9nnd_openshift-multus(ac51a893-4a0c-44ed-8284-0aac9e8d02ef)\"" pod="openshift-multus/multus-m9nnd" podUID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.433932 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:42 crc kubenswrapper[4730]: E0221 00:07:42.434059 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.434177 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.434227 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:42 crc kubenswrapper[4730]: I0221 00:07:42.433940 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:42 crc kubenswrapper[4730]: E0221 00:07:42.434413 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:42 crc kubenswrapper[4730]: E0221 00:07:42.434600 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:42 crc kubenswrapper[4730]: E0221 00:07:42.434734 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:43 crc kubenswrapper[4730]: I0221 00:07:43.243957 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/1.log" Feb 21 00:07:44 crc kubenswrapper[4730]: I0221 00:07:44.433885 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:44 crc kubenswrapper[4730]: I0221 00:07:44.433927 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:44 crc kubenswrapper[4730]: I0221 00:07:44.434115 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:44 crc kubenswrapper[4730]: E0221 00:07:44.434234 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:44 crc kubenswrapper[4730]: I0221 00:07:44.434325 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:44 crc kubenswrapper[4730]: E0221 00:07:44.434470 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:44 crc kubenswrapper[4730]: E0221 00:07:44.434812 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:44 crc kubenswrapper[4730]: E0221 00:07:44.434663 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:46 crc kubenswrapper[4730]: I0221 00:07:46.434518 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:46 crc kubenswrapper[4730]: I0221 00:07:46.434611 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:46 crc kubenswrapper[4730]: I0221 00:07:46.434645 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:46 crc kubenswrapper[4730]: I0221 00:07:46.434730 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:46 crc kubenswrapper[4730]: E0221 00:07:46.435021 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:46 crc kubenswrapper[4730]: E0221 00:07:46.435182 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:46 crc kubenswrapper[4730]: E0221 00:07:46.435403 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:46 crc kubenswrapper[4730]: E0221 00:07:46.435615 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:48 crc kubenswrapper[4730]: I0221 00:07:48.433933 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:48 crc kubenswrapper[4730]: I0221 00:07:48.434057 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:48 crc kubenswrapper[4730]: I0221 00:07:48.433951 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.434115 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:48 crc kubenswrapper[4730]: I0221 00:07:48.436610 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.436589 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.436820 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.436975 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.454868 4730 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Feb 21 00:07:48 crc kubenswrapper[4730]: E0221 00:07:48.555333 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:07:50 crc kubenswrapper[4730]: I0221 00:07:50.434246 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:50 crc kubenswrapper[4730]: I0221 00:07:50.434331 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:50 crc kubenswrapper[4730]: I0221 00:07:50.434364 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:50 crc kubenswrapper[4730]: E0221 00:07:50.434475 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:50 crc kubenswrapper[4730]: I0221 00:07:50.434504 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:50 crc kubenswrapper[4730]: E0221 00:07:50.434656 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:50 crc kubenswrapper[4730]: E0221 00:07:50.434857 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:50 crc kubenswrapper[4730]: E0221 00:07:50.435136 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:52 crc kubenswrapper[4730]: I0221 00:07:52.434612 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:52 crc kubenswrapper[4730]: I0221 00:07:52.434654 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:52 crc kubenswrapper[4730]: E0221 00:07:52.434838 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:52 crc kubenswrapper[4730]: I0221 00:07:52.435009 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:52 crc kubenswrapper[4730]: E0221 00:07:52.435156 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:52 crc kubenswrapper[4730]: I0221 00:07:52.435201 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:52 crc kubenswrapper[4730]: E0221 00:07:52.435587 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:52 crc kubenswrapper[4730]: E0221 00:07:52.435680 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:53 crc kubenswrapper[4730]: E0221 00:07:53.557543 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.435161 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.435255 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.436984 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.437096 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:54 crc kubenswrapper[4730]: E0221 00:07:54.437288 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:54 crc kubenswrapper[4730]: E0221 00:07:54.437448 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:54 crc kubenswrapper[4730]: E0221 00:07:54.437675 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.437765 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:07:54 crc kubenswrapper[4730]: E0221 00:07:54.438016 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:54 crc kubenswrapper[4730]: I0221 00:07:54.438158 4730 scope.go:117] "RemoveContainer" containerID="ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6" Feb 21 00:07:54 crc kubenswrapper[4730]: E0221 00:07:54.438175 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:07:55 crc kubenswrapper[4730]: I0221 00:07:55.301758 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/1.log" Feb 21 00:07:55 crc kubenswrapper[4730]: I0221 00:07:55.301838 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerStarted","Data":"ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432"} Feb 21 00:07:56 crc kubenswrapper[4730]: I0221 00:07:56.434455 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:56 crc kubenswrapper[4730]: E0221 00:07:56.434709 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:56 crc kubenswrapper[4730]: I0221 00:07:56.435049 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:56 crc kubenswrapper[4730]: E0221 00:07:56.435153 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:56 crc kubenswrapper[4730]: I0221 00:07:56.435254 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:56 crc kubenswrapper[4730]: I0221 00:07:56.435355 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:56 crc kubenswrapper[4730]: E0221 00:07:56.435537 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:56 crc kubenswrapper[4730]: E0221 00:07:56.435704 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:58 crc kubenswrapper[4730]: I0221 00:07:58.434545 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:07:58 crc kubenswrapper[4730]: I0221 00:07:58.435188 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:07:58 crc kubenswrapper[4730]: E0221 00:07:58.437096 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:07:58 crc kubenswrapper[4730]: I0221 00:07:58.437164 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:07:58 crc kubenswrapper[4730]: I0221 00:07:58.437128 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:07:58 crc kubenswrapper[4730]: E0221 00:07:58.437349 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:07:58 crc kubenswrapper[4730]: E0221 00:07:58.437476 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:07:58 crc kubenswrapper[4730]: E0221 00:07:58.437657 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:07:58 crc kubenswrapper[4730]: E0221 00:07:58.559020 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:08:00 crc kubenswrapper[4730]: I0221 00:08:00.434315 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:00 crc kubenswrapper[4730]: E0221 00:08:00.434967 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:00 crc kubenswrapper[4730]: I0221 00:08:00.434492 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:00 crc kubenswrapper[4730]: E0221 00:08:00.435068 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:00 crc kubenswrapper[4730]: I0221 00:08:00.434531 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:00 crc kubenswrapper[4730]: E0221 00:08:00.435125 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:00 crc kubenswrapper[4730]: I0221 00:08:00.434357 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:00 crc kubenswrapper[4730]: E0221 00:08:00.435168 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:02 crc kubenswrapper[4730]: I0221 00:08:02.434281 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:02 crc kubenswrapper[4730]: I0221 00:08:02.434324 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:02 crc kubenswrapper[4730]: E0221 00:08:02.434426 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:02 crc kubenswrapper[4730]: I0221 00:08:02.434494 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:02 crc kubenswrapper[4730]: I0221 00:08:02.434282 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:02 crc kubenswrapper[4730]: E0221 00:08:02.434636 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:02 crc kubenswrapper[4730]: E0221 00:08:02.434667 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:02 crc kubenswrapper[4730]: E0221 00:08:02.435452 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:03 crc kubenswrapper[4730]: E0221 00:08:03.560897 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:08:04 crc kubenswrapper[4730]: I0221 00:08:04.434639 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:04 crc kubenswrapper[4730]: I0221 00:08:04.434684 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:04 crc kubenswrapper[4730]: E0221 00:08:04.434849 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:04 crc kubenswrapper[4730]: I0221 00:08:04.434937 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:04 crc kubenswrapper[4730]: I0221 00:08:04.434973 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:04 crc kubenswrapper[4730]: E0221 00:08:04.435126 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:04 crc kubenswrapper[4730]: E0221 00:08:04.435388 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:04 crc kubenswrapper[4730]: E0221 00:08:04.435535 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:05 crc kubenswrapper[4730]: I0221 00:08:05.435764 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:08:05 crc kubenswrapper[4730]: E0221 00:08:05.437031 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-rjptr_openshift-ovn-kubernetes(cfb5b15d-d281-4633-9930-905d05ed76e8)\"" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" Feb 21 00:08:06 crc kubenswrapper[4730]: I0221 00:08:06.434786 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:06 crc kubenswrapper[4730]: I0221 00:08:06.434843 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:06 crc kubenswrapper[4730]: I0221 00:08:06.434874 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:06 crc kubenswrapper[4730]: E0221 00:08:06.435023 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:06 crc kubenswrapper[4730]: I0221 00:08:06.435068 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:06 crc kubenswrapper[4730]: E0221 00:08:06.435231 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:06 crc kubenswrapper[4730]: E0221 00:08:06.435508 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:06 crc kubenswrapper[4730]: E0221 00:08:06.435604 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:08 crc kubenswrapper[4730]: I0221 00:08:08.434209 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:08 crc kubenswrapper[4730]: I0221 00:08:08.434336 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:08 crc kubenswrapper[4730]: I0221 00:08:08.434436 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:08 crc kubenswrapper[4730]: I0221 00:08:08.436565 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:08 crc kubenswrapper[4730]: E0221 00:08:08.436557 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:08 crc kubenswrapper[4730]: E0221 00:08:08.436712 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:08 crc kubenswrapper[4730]: E0221 00:08:08.436888 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:08 crc kubenswrapper[4730]: E0221 00:08:08.437158 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:08 crc kubenswrapper[4730]: E0221 00:08:08.562147 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:08:10 crc kubenswrapper[4730]: I0221 00:08:10.433832 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:10 crc kubenswrapper[4730]: I0221 00:08:10.433956 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:10 crc kubenswrapper[4730]: I0221 00:08:10.433987 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:10 crc kubenswrapper[4730]: I0221 00:08:10.433838 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:10 crc kubenswrapper[4730]: E0221 00:08:10.434072 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:10 crc kubenswrapper[4730]: E0221 00:08:10.434193 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:10 crc kubenswrapper[4730]: E0221 00:08:10.434382 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:10 crc kubenswrapper[4730]: E0221 00:08:10.434616 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:12 crc kubenswrapper[4730]: I0221 00:08:12.434184 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:12 crc kubenswrapper[4730]: I0221 00:08:12.434561 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:12 crc kubenswrapper[4730]: E0221 00:08:12.434655 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:12 crc kubenswrapper[4730]: E0221 00:08:12.434815 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:12 crc kubenswrapper[4730]: I0221 00:08:12.434408 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:12 crc kubenswrapper[4730]: E0221 00:08:12.435008 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:12 crc kubenswrapper[4730]: I0221 00:08:12.434336 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:12 crc kubenswrapper[4730]: E0221 00:08:12.435166 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:13 crc kubenswrapper[4730]: E0221 00:08:13.563609 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:08:14 crc kubenswrapper[4730]: I0221 00:08:14.435045 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:14 crc kubenswrapper[4730]: I0221 00:08:14.435186 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:14 crc kubenswrapper[4730]: I0221 00:08:14.435153 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:14 crc kubenswrapper[4730]: I0221 00:08:14.435124 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:14 crc kubenswrapper[4730]: E0221 00:08:14.435392 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:14 crc kubenswrapper[4730]: E0221 00:08:14.435586 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:14 crc kubenswrapper[4730]: E0221 00:08:14.435967 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:14 crc kubenswrapper[4730]: E0221 00:08:14.436070 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.345642 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.345731 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.345967 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346008 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346033 4730 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.345967 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346134 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-02-21 00:10:18.346104875 +0000 UTC m=+270.636940645 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346150 4730 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346175 4730 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.346243 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-02-21 00:10:18.346221348 +0000 UTC m=+270.637057128 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.434768 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.434833 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.435010 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.435045 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.435092 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.435246 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.435452 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.435611 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.446352 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.446600 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:10:18.44655419 +0000 UTC m=+270.737390030 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.547661 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:16 crc kubenswrapper[4730]: I0221 00:08:16.547760 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.547801 4730 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.547869 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:10:18.54785125 +0000 UTC m=+270.838686990 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.548026 4730 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:08:16 crc kubenswrapper[4730]: E0221 00:08:16.548164 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-02-21 00:10:18.548133068 +0000 UTC m=+270.838968838 (durationBeforeRetry 2m2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Feb 21 00:08:18 crc kubenswrapper[4730]: I0221 00:08:18.434602 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:18 crc kubenswrapper[4730]: I0221 00:08:18.434851 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:18 crc kubenswrapper[4730]: I0221 00:08:18.434861 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:18 crc kubenswrapper[4730]: E0221 00:08:18.437318 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:18 crc kubenswrapper[4730]: I0221 00:08:18.437351 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:18 crc kubenswrapper[4730]: E0221 00:08:18.437578 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:18 crc kubenswrapper[4730]: E0221 00:08:18.438184 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:18 crc kubenswrapper[4730]: E0221 00:08:18.438310 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:18 crc kubenswrapper[4730]: I0221 00:08:18.438665 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:08:18 crc kubenswrapper[4730]: E0221 00:08:18.564401 4730 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Feb 21 00:08:19 crc kubenswrapper[4730]: I0221 00:08:19.382440 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krrw8"] Feb 21 00:08:19 crc kubenswrapper[4730]: I0221 00:08:19.407057 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/3.log" Feb 21 00:08:19 crc kubenswrapper[4730]: I0221 00:08:19.409474 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerStarted","Data":"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8"} Feb 21 00:08:19 crc kubenswrapper[4730]: I0221 00:08:19.409508 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:19 crc kubenswrapper[4730]: E0221 00:08:19.409615 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:19 crc kubenswrapper[4730]: I0221 00:08:19.409843 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:08:20 crc kubenswrapper[4730]: I0221 00:08:20.434831 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:20 crc kubenswrapper[4730]: I0221 00:08:20.434887 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:20 crc kubenswrapper[4730]: I0221 00:08:20.434953 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:20 crc kubenswrapper[4730]: E0221 00:08:20.435093 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:20 crc kubenswrapper[4730]: E0221 00:08:20.435212 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:20 crc kubenswrapper[4730]: E0221 00:08:20.435324 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:21 crc kubenswrapper[4730]: I0221 00:08:21.434706 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:21 crc kubenswrapper[4730]: E0221 00:08:21.434943 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:22 crc kubenswrapper[4730]: I0221 00:08:22.434577 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:22 crc kubenswrapper[4730]: E0221 00:08:22.434784 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Feb 21 00:08:22 crc kubenswrapper[4730]: I0221 00:08:22.434584 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:22 crc kubenswrapper[4730]: I0221 00:08:22.434880 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:22 crc kubenswrapper[4730]: E0221 00:08:22.435103 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:08:22 crc kubenswrapper[4730]: E0221 00:08:22.435277 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Feb 21 00:08:23 crc kubenswrapper[4730]: I0221 00:08:23.434969 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:23 crc kubenswrapper[4730]: E0221 00:08:23.435600 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-krrw8" podUID="73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971" Feb 21 00:08:23 crc kubenswrapper[4730]: I0221 00:08:23.441768 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.434391 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.434488 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.434978 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.444716 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.444861 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.445131 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 21 00:08:24 crc kubenswrapper[4730]: I0221 00:08:24.445487 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 21 00:08:25 crc kubenswrapper[4730]: I0221 00:08:25.433976 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:25 crc kubenswrapper[4730]: I0221 00:08:25.436795 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 21 00:08:25 crc kubenswrapper[4730]: I0221 00:08:25.436959 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.871469 4730 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.922527 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.923604 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.923711 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.924635 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.928046 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.928132 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.928500 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.929181 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929344 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929381 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929355 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929413 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929400 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929447 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929391 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929479 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4": failed to list *v1.Secret: secrets "machine-approver-sa-dockercfg-nl2j4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929494 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929529 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929436 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929513 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-sa-dockercfg-nl2j4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-sa-dockercfg-nl2j4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929598 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929605 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-tls": failed to list *v1.Secret: secrets "machine-approver-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929619 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.929666 4730 reflector.go:561] object-"openshift-cluster-machine-approver"/"machine-approver-config": failed to list *v1.ConfigMap: configmaps "machine-approver-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-machine-approver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929678 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929633 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-approver-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929475 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.929693 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-machine-approver\"/\"machine-approver-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-approver-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-machine-approver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.932107 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.932847 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.935022 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ssd6c"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.935480 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.936726 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.937140 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.937383 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.938204 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.938553 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.938597 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.938722 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.959814 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: W0221 00:08:26.960553 4730 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Feb 21 00:08:26 crc kubenswrapper[4730]: E0221 00:08:26.960622 4730 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.960789 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.961818 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962794 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962851 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962886 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962896 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962938 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.962979 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.963259 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.966023 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.967933 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.968560 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.968556 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.968647 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.968752 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.963328 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.973030 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m9mhf"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.975318 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.976625 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.977178 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.983794 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.987575 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.988554 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.988867 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989217 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989410 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989641 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989826 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989928 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.990077 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.990302 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.989870 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.990984 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29527200-hgrnz"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.991445 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f5csk"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.991659 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.991734 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.992007 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.992811 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.992855 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.993211 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.994232 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 21 00:08:26 crc kubenswrapper[4730]: I0221 00:08:26.994314 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.000037 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bsnm7"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.000363 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.000553 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.000933 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.000986 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001315 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001321 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001561 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001675 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001851 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001877 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.001965 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002127 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002218 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002280 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002286 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ht8kr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002786 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.002955 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.011046 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.011934 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.022390 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5rg5n"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.023322 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-4wrmp"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.023698 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024010 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024499 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-images\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024529 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024546 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024671 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024693 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rmq4n"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024721 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024757 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024787 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrkx6\" (UniqueName: \"kubernetes.io/projected/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-kube-api-access-rrkx6\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024823 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024849 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/52967242-12a0-48ed-a278-791b2cb76dc7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024870 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlw5v\" (UniqueName: \"kubernetes.io/projected/05d038d6-7946-445b-a1d3-c2d6f0044b7b-kube-api-access-qlw5v\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024893 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52967242-12a0-48ed-a278-791b2cb76dc7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.024941 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ll5bt\" (UniqueName: \"kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025151 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025243 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025289 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025432 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025581 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d038d6-7946-445b-a1d3-c2d6f0044b7b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025624 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025663 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025693 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025720 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025750 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp59f\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-kube-api-access-cp59f\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025779 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025812 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025835 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025924 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-proxy-tls\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.025957 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.041695 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.046490 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.047386 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.047931 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.048154 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.065460 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.065765 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.065925 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.066057 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.066354 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.066516 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.048296 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.068199 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.068616 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.068956 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.069518 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.069656 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.069754 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.071096 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.071433 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.071621 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074306 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074450 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m9sb5"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074481 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074561 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074692 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.074761 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-twxxw"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.075086 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.075256 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.075352 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.078638 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.078983 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.079172 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.079383 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.079594 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.080158 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.080772 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.080963 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081176 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081295 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081530 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081648 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081749 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.081874 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.082772 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.083034 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.083182 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.083825 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.085561 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.094234 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.094800 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.095282 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-bk7rn"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.095774 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.096842 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-444fb"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.097270 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.098722 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-mk5nk"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.099193 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.099207 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.099692 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.101267 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.101838 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.101933 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.116764 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.117186 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.117641 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.118175 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.118172 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.118194 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.113467 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.118225 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.113828 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.113987 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.121499 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jpj6l"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.122327 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.122925 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.122344 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.124139 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.124541 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.125471 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.125854 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.126578 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.126777 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.127239 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.128971 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.129344 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.129677 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130038 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130343 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-client\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130371 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130390 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-cabundle\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130406 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2c7c\" (UniqueName: \"kubernetes.io/projected/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-kube-api-access-r2c7c\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130424 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-client\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130439 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-service-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130462 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130478 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8jt6\" (UniqueName: \"kubernetes.io/projected/98524d44-741a-4550-b185-5116a1498602-kube-api-access-v8jt6\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130494 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130510 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130524 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1012895-d7ed-4a60-95fa-8521f1e57b3d-proxy-tls\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130542 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/52967242-12a0-48ed-a278-791b2cb76dc7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130560 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlw5v\" (UniqueName: \"kubernetes.io/projected/05d038d6-7946-445b-a1d3-c2d6f0044b7b-kube-api-access-qlw5v\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130577 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130594 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130610 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-config\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130627 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130644 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxldr\" (UniqueName: \"kubernetes.io/projected/339c3294-8719-4942-9998-67605f596945-kube-api-access-qxldr\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130661 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52967242-12a0-48ed-a278-791b2cb76dc7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130676 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.130692 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131042 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ll5bt\" (UniqueName: \"kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131058 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131073 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tv8x\" (UniqueName: \"kubernetes.io/projected/2d012ccd-e3fb-4dbb-b375-e70f9d803919-kube-api-access-7tv8x\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131088 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpskf\" (UniqueName: \"kubernetes.io/projected/f1012895-d7ed-4a60-95fa-8521f1e57b3d-kube-api-access-cpskf\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131104 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-metrics-certs\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131117 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98524d44-741a-4550-b185-5116a1498602-serving-cert\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131133 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131148 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6q4j\" (UniqueName: \"kubernetes.io/projected/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-kube-api-access-w6q4j\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131162 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-node-pullsecrets\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131177 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131191 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131208 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131237 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f98bd\" (UniqueName: \"kubernetes.io/projected/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-kube-api-access-f98bd\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131251 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e921b9fb-4785-4303-acd0-f04a0b5c4da8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131266 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1012895-d7ed-4a60-95fa-8521f1e57b3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131281 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86h2s\" (UniqueName: \"kubernetes.io/projected/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-kube-api-access-86h2s\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131297 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131312 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-trusted-ca\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131327 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c79c4da-0098-43d4-83c9-0e140f37c8fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131341 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-trusted-ca-bundle\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131355 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrcxk\" (UniqueName: \"kubernetes.io/projected/f28f3f0c-4af6-4f64-baef-b2fdff773582-kube-api-access-xrcxk\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131372 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d038d6-7946-445b-a1d3-c2d6f0044b7b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131377 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29527200-hgrnz"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131386 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-key\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131403 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131417 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7965\" (UniqueName: \"kubernetes.io/projected/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-kube-api-access-p7965\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131432 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131447 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-image-import-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131460 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e236e2c-f743-4aac-8b58-f9dc13769abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131477 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131492 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-serving-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131506 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131520 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131535 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-client\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131549 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-encryption-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131563 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8sj2\" (UniqueName: \"kubernetes.io/projected/2c79c4da-0098-43d4-83c9-0e140f37c8fe-kube-api-access-v8sj2\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131580 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131593 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-service-ca\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131616 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131631 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp59f\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-kube-api-access-cp59f\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131645 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c79c4da-0098-43d4-83c9-0e140f37c8fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131660 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-serving-cert\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131674 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131688 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit-dir\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131704 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrr6d\" (UniqueName: \"kubernetes.io/projected/0d89fc71-51e8-4595-b9d4-fc0cd8103017-kube-api-access-jrr6d\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131724 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-serving-cert\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131737 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klnpr\" (UniqueName: \"kubernetes.io/projected/2eb73c39-8598-4fdf-afe6-314a98db6331-kube-api-access-klnpr\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131751 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rkn8f\" (UniqueName: \"kubernetes.io/projected/e921b9fb-4785-4303-acd0-f04a0b5c4da8-kube-api-access-rkn8f\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131764 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-oauth-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131780 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131793 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131809 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131824 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-serving-cert\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131838 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-dir\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131852 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4szsc\" (UniqueName: \"kubernetes.io/projected/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-kube-api-access-4szsc\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131882 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-proxy-tls\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131896 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svb4m\" (UniqueName: \"kubernetes.io/projected/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-kube-api-access-svb4m\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131926 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e921b9fb-4785-4303-acd0-f04a0b5c4da8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131941 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-default-certificate\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131956 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131971 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131985 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.131998 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/339c3294-8719-4942-9998-67605f596945-metrics-tls\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132013 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb73c39-8598-4fdf-afe6-314a98db6331-serving-cert\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132037 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-config\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132052 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132067 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d89fc71-51e8-4595-b9d4-fc0cd8103017-service-ca-bundle\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132081 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-stats-auth\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132096 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e236e2c-f743-4aac-8b58-f9dc13769abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132111 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132125 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-oauth-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132141 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-config\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132164 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-images\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132181 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-service-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132196 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132212 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-config\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132227 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132242 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132256 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpcjv\" (UniqueName: \"kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132272 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132288 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132305 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jgssk\" (UniqueName: \"kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132319 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132337 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132353 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxxnf\" (UniqueName: \"kubernetes.io/projected/49975767-e31f-4b1b-9fbb-0e63abb8cd47-kube-api-access-lxxnf\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132369 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132384 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132398 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132412 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-srv-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132428 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrkx6\" (UniqueName: \"kubernetes.io/projected/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-kube-api-access-rrkx6\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132443 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-serving-cert\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132458 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8285\" (UniqueName: \"kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.132474 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.133153 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.134140 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ssd6c"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.151798 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/52967242-12a0-48ed-a278-791b2cb76dc7-trusted-ca\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.152300 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.153054 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-images\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.153729 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.153932 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.154173 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.154679 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m9mhf"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.156252 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-proxy-tls\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.157597 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.157833 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.157872 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.159025 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.159853 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/05d038d6-7946-445b-a1d3-c2d6f0044b7b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.160477 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.161542 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.161645 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.162145 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.162308 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/52967242-12a0-48ed-a278-791b2cb76dc7-metrics-tls\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.163283 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-2q9tl"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.163552 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.163924 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.164250 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.166094 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-lpmdh"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.167070 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.169731 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.169788 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f5csk"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.176805 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.178818 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.179810 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.180240 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mk5nk"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.181290 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ht8kr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.182565 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.183883 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bsnm7"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.184766 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.185302 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.186374 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.187746 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.188709 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rmq4n"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.189864 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5rg5n"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.191305 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.192277 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.193361 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m9sb5"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.194434 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.195518 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.196482 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.197464 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.199736 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.201299 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bk7rn"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.202312 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-twxxw"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.203757 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.205003 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.205537 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.206617 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-444fb"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.207626 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.208662 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.210018 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.210989 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.212397 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jpj6l"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.213475 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lpmdh"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.214887 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-djs5z"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.215774 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.215897 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n66hq"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.216823 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.217079 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.218399 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.219360 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.220510 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-djs5z"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.221584 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n66hq"] Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.225985 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233179 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-srv-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233238 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233273 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233298 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233327 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-serving-cert\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233349 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8285\" (UniqueName: \"kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233371 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233424 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-cabundle\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233454 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-client\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233502 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2c7c\" (UniqueName: \"kubernetes.io/projected/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-kube-api-access-r2c7c\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233540 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-client\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233569 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-service-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233593 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8jt6\" (UniqueName: \"kubernetes.io/projected/98524d44-741a-4550-b185-5116a1498602-kube-api-access-v8jt6\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233620 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233641 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233665 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1012895-d7ed-4a60-95fa-8521f1e57b3d-proxy-tls\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233695 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233717 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233741 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxldr\" (UniqueName: \"kubernetes.io/projected/339c3294-8719-4942-9998-67605f596945-kube-api-access-qxldr\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233766 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-config\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233788 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233811 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233833 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233855 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-metrics-certs\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233875 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98524d44-741a-4550-b185-5116a1498602-serving-cert\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233944 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233967 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tv8x\" (UniqueName: \"kubernetes.io/projected/2d012ccd-e3fb-4dbb-b375-e70f9d803919-kube-api-access-7tv8x\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.233988 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpskf\" (UniqueName: \"kubernetes.io/projected/f1012895-d7ed-4a60-95fa-8521f1e57b3d-kube-api-access-cpskf\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234009 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6q4j\" (UniqueName: \"kubernetes.io/projected/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-kube-api-access-w6q4j\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234029 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-node-pullsecrets\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234054 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234077 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234100 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234128 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f98bd\" (UniqueName: \"kubernetes.io/projected/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-kube-api-access-f98bd\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234155 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e921b9fb-4785-4303-acd0-f04a0b5c4da8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234181 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86h2s\" (UniqueName: \"kubernetes.io/projected/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-kube-api-access-86h2s\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234208 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1012895-d7ed-4a60-95fa-8521f1e57b3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234233 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrcxk\" (UniqueName: \"kubernetes.io/projected/f28f3f0c-4af6-4f64-baef-b2fdff773582-kube-api-access-xrcxk\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234260 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-trusted-ca\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234286 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c79c4da-0098-43d4-83c9-0e140f37c8fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234307 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-trusted-ca-bundle\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234331 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-key\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234352 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234374 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7965\" (UniqueName: \"kubernetes.io/projected/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-kube-api-access-p7965\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234398 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-image-import-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234421 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e236e2c-f743-4aac-8b58-f9dc13769abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234450 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234481 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-serving-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234505 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8sj2\" (UniqueName: \"kubernetes.io/projected/2c79c4da-0098-43d4-83c9-0e140f37c8fe-kube-api-access-v8sj2\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234531 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234551 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-client\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234571 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-encryption-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234633 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-service-ca\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234654 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-serving-cert\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234691 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c79c4da-0098-43d4-83c9-0e140f37c8fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234723 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit-dir\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234745 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrr6d\" (UniqueName: \"kubernetes.io/projected/0d89fc71-51e8-4595-b9d4-fc0cd8103017-kube-api-access-jrr6d\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234767 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-serving-cert\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234787 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klnpr\" (UniqueName: \"kubernetes.io/projected/2eb73c39-8598-4fdf-afe6-314a98db6331-kube-api-access-klnpr\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234810 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rkn8f\" (UniqueName: \"kubernetes.io/projected/e921b9fb-4785-4303-acd0-f04a0b5c4da8-kube-api-access-rkn8f\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234830 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-oauth-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234852 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4szsc\" (UniqueName: \"kubernetes.io/projected/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-kube-api-access-4szsc\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234890 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234932 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-serving-cert\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234955 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-dir\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.234977 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svb4m\" (UniqueName: \"kubernetes.io/projected/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-kube-api-access-svb4m\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235010 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-default-certificate\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235031 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e921b9fb-4785-4303-acd0-f04a0b5c4da8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235054 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235078 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/339c3294-8719-4942-9998-67605f596945-metrics-tls\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235104 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235142 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-config\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235164 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb73c39-8598-4fdf-afe6-314a98db6331-serving-cert\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235188 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235210 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-oauth-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235233 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235256 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d89fc71-51e8-4595-b9d4-fc0cd8103017-service-ca-bundle\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235261 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235279 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-stats-auth\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235303 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e236e2c-f743-4aac-8b58-f9dc13769abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235347 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-config\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235369 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235403 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-service-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235429 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235459 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-config\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235487 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235516 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpcjv\" (UniqueName: \"kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235552 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jgssk\" (UniqueName: \"kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235562 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-trusted-ca\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235581 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235640 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxxnf\" (UniqueName: \"kubernetes.io/projected/49975767-e31f-4b1b-9fbb-0e63abb8cd47-kube-api-access-lxxnf\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.235969 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.236474 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f1012895-d7ed-4a60-95fa-8521f1e57b3d-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.237266 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-node-pullsecrets\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.237500 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.237850 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit-dir\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.238088 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-dir\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.238580 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-serving-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.238676 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-image-import-ca\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.238696 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-metrics-certs\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.239377 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-trusted-ca-bundle\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.239922 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240036 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2d012ccd-e3fb-4dbb-b375-e70f9d803919-audit\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240364 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240582 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-serving-cert\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240777 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240938 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.240963 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-config\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.241045 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2eb73c39-8598-4fdf-afe6-314a98db6331-service-ca-bundle\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.241252 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.241744 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98524d44-741a-4550-b185-5116a1498602-config\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.241839 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.241977 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-encryption-config\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.242231 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-client\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.242333 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98524d44-741a-4550-b185-5116a1498602-serving-cert\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.242348 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-serving-cert\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.242558 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.242783 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0d89fc71-51e8-4595-b9d4-fc0cd8103017-service-ca-bundle\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.243477 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-config\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.243573 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-service-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.243692 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.243738 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2eb73c39-8598-4fdf-afe6-314a98db6331-serving-cert\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.244019 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/f1012895-d7ed-4a60-95fa-8521f1e57b3d-proxy-tls\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.244066 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/339c3294-8719-4942-9998-67605f596945-metrics-tls\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.245067 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-ca\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.245462 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.246195 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-default-certificate\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.246571 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-serving-cert\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.248527 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-etcd-client\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.248536 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2d012ccd-e3fb-4dbb-b375-e70f9d803919-etcd-client\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.250806 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.251333 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0d89fc71-51e8-4595-b9d4-fc0cd8103017-stats-auth\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.265947 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.285569 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.307372 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.319235 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-srv-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.325953 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.334036 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.346368 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.366121 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.386321 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.408048 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.409629 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-config\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.426795 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.447091 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.467083 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.471692 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-serving-cert\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.486463 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.506281 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.527746 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.546517 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.552158 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-key\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.566780 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.569323 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/f28f3f0c-4af6-4f64-baef-b2fdff773582-signing-cabundle\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.586805 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.627419 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.647103 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.668036 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.687437 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.708454 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.726276 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.748000 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.766833 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.787830 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.806674 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.827297 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.842398 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-oauth-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.847424 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.875217 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.887020 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.893429 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.906330 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.927015 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.946029 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.949319 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-service-ca\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.967518 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.968854 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-console-config\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.985842 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 21 00:08:27 crc kubenswrapper[4730]: I0221 00:08:27.990614 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-oauth-serving-cert\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.015849 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.018279 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-trusted-ca-bundle\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.025946 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.032278 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e921b9fb-4785-4303-acd0-f04a0b5c4da8-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.046186 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.066071 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.077162 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e921b9fb-4785-4303-acd0-f04a0b5c4da8-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.086398 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.105819 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.124226 4730 request.go:700] Waited for 1.005558662s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator-operator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.125415 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.146474 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.167897 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.171636 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2c79c4da-0098-43d4-83c9-0e140f37c8fe-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.186305 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.205879 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.207300 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2c79c4da-0098-43d4-83c9-0e140f37c8fe-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.226807 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.233751 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6e236e2c-f743-4aac-8b58-f9dc13769abb-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.233931 4730 configmap.go:193] Couldn't get configMap openshift-kube-apiserver-operator/kube-apiserver-operator-config: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.233958 4730 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.234015 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config podName:6e236e2c-f743-4aac-8b58-f9dc13769abb nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.733990821 +0000 UTC m=+161.024826571 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config") pod "kube-apiserver-operator-766d6c64bb-vxwd9" (UID: "6e236e2c-f743-4aac-8b58-f9dc13769abb") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.234039 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca podName:6f272e98-1b2b-4c55-b095-5f0b2a2f8972 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.734026232 +0000 UTC m=+161.024861972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca") pod "apiserver-7bbb656c7d-zwmzp" (UID: "6f272e98-1b2b-4c55-b095-5f0b2a2f8972") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.237711 4730 secret.go:188] Couldn't get secret openshift-cluster-machine-approver/machine-approver-tls: failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.237789 4730 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/machine-approver-config: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.237726 4730 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.237741 4730 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.237870 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls podName:6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.737827553 +0000 UTC m=+161.028663383 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-approver-tls" (UniqueName: "kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls") pod "machine-approver-56656f9798-5gb7h" (UID: "6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1") : failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238067 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config podName:6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.737993488 +0000 UTC m=+161.028829258 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config") pod "machine-approver-56656f9798-5gb7h" (UID: "6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238111 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config podName:6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.738095831 +0000 UTC m=+161.028931591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config") pod "machine-approver-56656f9798-5gb7h" (UID: "6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238151 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config podName:6f272e98-1b2b-4c55-b095-5f0b2a2f8972 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.738135842 +0000 UTC m=+161.028971823 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config") pod "apiserver-7bbb656c7d-zwmzp" (UID: "6f272e98-1b2b-4c55-b095-5f0b2a2f8972") : failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238164 4730 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238236 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies podName:6f272e98-1b2b-4c55-b095-5f0b2a2f8972 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.738212025 +0000 UTC m=+161.029048015 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies") pod "apiserver-7bbb656c7d-zwmzp" (UID: "6f272e98-1b2b-4c55-b095-5f0b2a2f8972") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238976 4730 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.238980 4730 secret.go:188] Couldn't get secret openshift-multus/multus-admission-controller-secret: failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.239036 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle podName:6f272e98-1b2b-4c55-b095-5f0b2a2f8972 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.739015598 +0000 UTC m=+161.029851548 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle") pod "apiserver-7bbb656c7d-zwmzp" (UID: "6f272e98-1b2b-4c55-b095-5f0b2a2f8972") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: E0221 00:08:28.239094 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs podName:49975767-e31f-4b1b-9fbb-0e63abb8cd47 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:28.73907048 +0000 UTC m=+161.029906430 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs") pod "multus-admission-controller-857f4d67dd-jpj6l" (UID: "49975767-e31f-4b1b-9fbb-0e63abb8cd47") : failed to sync secret cache: timed out waiting for the condition Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.247315 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.265891 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.286453 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.306562 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.326117 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.347682 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.366348 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.386602 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.406662 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.426212 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.446807 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.467753 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.486181 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.506540 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.526533 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.546829 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.566471 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.586297 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.606244 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.633182 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.646726 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.686493 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.706471 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.753627 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp59f\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-kube-api-access-cp59f\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.762644 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.762778 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.762922 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763002 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763112 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763193 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763275 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763348 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.763482 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.764071 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e236e2c-f743-4aac-8b58-f9dc13769abb-config\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.765968 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/49975767-e31f-4b1b-9fbb-0e63abb8cd47-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.766269 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.796191 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlw5v\" (UniqueName: \"kubernetes.io/projected/05d038d6-7946-445b-a1d3-c2d6f0044b7b-kube-api-access-qlw5v\") pod \"cluster-samples-operator-665b6dd947-952vl\" (UID: \"05d038d6-7946-445b-a1d3-c2d6f0044b7b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.813534 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ll5bt\" (UniqueName: \"kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt\") pod \"oauth-openshift-558db77b4-ssd6c\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.826135 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/52967242-12a0-48ed-a278-791b2cb76dc7-bound-sa-token\") pod \"ingress-operator-5b745b69d9-lpf96\" (UID: \"52967242-12a0-48ed-a278-791b2cb76dc7\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.826605 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.855554 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.861607 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrkx6\" (UniqueName: \"kubernetes.io/projected/a132f6d0-9acf-4dd2-a773-2ae6640dae8b-kube-api-access-rrkx6\") pod \"machine-config-operator-74547568cd-8fj4m\" (UID: \"a132f6d0-9acf-4dd2-a773-2ae6640dae8b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.870243 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.885604 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.905635 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.925704 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.946289 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.967767 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.986778 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 21 00:08:28 crc kubenswrapper[4730]: I0221 00:08:28.995342 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.000672 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.006812 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.026126 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.047200 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.065701 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.085695 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.091153 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.113875 4730 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.144861 4730 request.go:700] Waited for 1.909409322s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/console/token Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.167993 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ssd6c"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.170357 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86h2s\" (UniqueName: \"kubernetes.io/projected/d34ad48a-e9a9-47dc-a4b2-16edb864a63a-kube-api-access-86h2s\") pod \"console-f9d7485db-mk5nk\" (UID: \"d34ad48a-e9a9-47dc-a4b2-16edb864a63a\") " pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.216686 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxxnf\" (UniqueName: \"kubernetes.io/projected/49975767-e31f-4b1b-9fbb-0e63abb8cd47-kube-api-access-lxxnf\") pod \"multus-admission-controller-857f4d67dd-jpj6l\" (UID: \"49975767-e31f-4b1b-9fbb-0e63abb8cd47\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.217783 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.220018 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tv8x\" (UniqueName: \"kubernetes.io/projected/2d012ccd-e3fb-4dbb-b375-e70f9d803919-kube-api-access-7tv8x\") pod \"apiserver-76f77b778f-m9mhf\" (UID: \"2d012ccd-e3fb-4dbb-b375-e70f9d803919\") " pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.238355 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.241694 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpskf\" (UniqueName: \"kubernetes.io/projected/f1012895-d7ed-4a60-95fa-8521f1e57b3d-kube-api-access-cpskf\") pod \"machine-config-controller-84d6567774-dtdxb\" (UID: \"f1012895-d7ed-4a60-95fa-8521f1e57b3d\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:29 crc kubenswrapper[4730]: W0221 00:08:29.249984 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod52967242_12a0_48ed_a278_791b2cb76dc7.slice/crio-aac9f14f67618216a57568846d813f5d35933140aac83ca4f462eba3cf993c20 WatchSource:0}: Error finding container aac9f14f67618216a57568846d813f5d35933140aac83ca4f462eba3cf993c20: Status 404 returned error can't find the container with id aac9f14f67618216a57568846d813f5d35933140aac83ca4f462eba3cf993c20 Feb 21 00:08:29 crc kubenswrapper[4730]: W0221 00:08:29.250799 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda132f6d0_9acf_4dd2_a773_2ae6640dae8b.slice/crio-2097f64cc1f4c02caeafffda5520e70d87d2a6c915fb819cf33306d8a929d1f7 WatchSource:0}: Error finding container 2097f64cc1f4c02caeafffda5520e70d87d2a6c915fb819cf33306d8a929d1f7: Status 404 returned error can't find the container with id 2097f64cc1f4c02caeafffda5520e70d87d2a6c915fb819cf33306d8a929d1f7 Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.260585 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrcxk\" (UniqueName: \"kubernetes.io/projected/f28f3f0c-4af6-4f64-baef-b2fdff773582-kube-api-access-xrcxk\") pod \"service-ca-9c57cc56f-m9sb5\" (UID: \"f28f3f0c-4af6-4f64-baef-b2fdff773582\") " pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.278112 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.279395 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrr6d\" (UniqueName: \"kubernetes.io/projected/0d89fc71-51e8-4595-b9d4-fc0cd8103017-kube-api-access-jrr6d\") pod \"router-default-5444994796-4wrmp\" (UID: \"0d89fc71-51e8-4595-b9d4-fc0cd8103017\") " pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.299250 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8285\" (UniqueName: \"kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285\") pod \"image-pruner-29527200-hgrnz\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.317535 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.319697 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8sj2\" (UniqueName: \"kubernetes.io/projected/2c79c4da-0098-43d4-83c9-0e140f37c8fe-kube-api-access-v8sj2\") pod \"kube-storage-version-migrator-operator-b67b599dd-t246w\" (UID: \"2c79c4da-0098-43d4-83c9-0e140f37c8fe\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.336091 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.341675 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7965\" (UniqueName: \"kubernetes.io/projected/34eac1cb-c6b6-441b-bb23-f0b95fea55f2-kube-api-access-p7965\") pod \"olm-operator-6b444d44fb-z6jtl\" (UID: \"34eac1cb-c6b6-441b-bb23-f0b95fea55f2\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.360279 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.363008 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2c7c\" (UniqueName: \"kubernetes.io/projected/07a4c1db-3aad-4085-9eed-cd05bb34d6f2-kube-api-access-r2c7c\") pod \"openshift-apiserver-operator-796bbdcf4f-vctw7\" (UID: \"07a4c1db-3aad-4085-9eed-cd05bb34d6f2\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.379399 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.386862 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rkn8f\" (UniqueName: \"kubernetes.io/projected/e921b9fb-4785-4303-acd0-f04a0b5c4da8-kube-api-access-rkn8f\") pod \"openshift-controller-manager-operator-756b6f6bc6-rt9q4\" (UID: \"e921b9fb-4785-4303-acd0-f04a0b5c4da8\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.400316 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svb4m\" (UniqueName: \"kubernetes.io/projected/6e025d1e-4c07-4cd2-8cb2-a34ae7641e16-kube-api-access-svb4m\") pod \"etcd-operator-b45778765-bsnm7\" (UID: \"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16\") " pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.404561 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.428209 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.431829 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jgssk\" (UniqueName: \"kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk\") pod \"route-controller-manager-6576b87f9c-lj2pw\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.440579 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.447030 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpcjv\" (UniqueName: \"kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv\") pod \"controller-manager-879f6c89f-2qjdj\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.449467 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.455923 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4wrmp" event={"ID":"0d89fc71-51e8-4595-b9d4-fc0cd8103017","Type":"ContainerStarted","Data":"bb50c99dde11b5a30811ae777d009823d4dc72e3905bb9579e2b8d2c4f295031"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.470416 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klnpr\" (UniqueName: \"kubernetes.io/projected/2eb73c39-8598-4fdf-afe6-314a98db6331-kube-api-access-klnpr\") pod \"authentication-operator-69f744f599-f5csk\" (UID: \"2eb73c39-8598-4fdf-afe6-314a98db6331\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.474856 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" event={"ID":"a132f6d0-9acf-4dd2-a773-2ae6640dae8b","Type":"ContainerStarted","Data":"e4a3bd0a27526dda35d6b4d8094ece11245db758a2b412427d76639dcf4020f4"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.474895 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" event={"ID":"a132f6d0-9acf-4dd2-a773-2ae6640dae8b","Type":"ContainerStarted","Data":"2097f64cc1f4c02caeafffda5520e70d87d2a6c915fb819cf33306d8a929d1f7"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.482998 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" event={"ID":"52967242-12a0-48ed-a278-791b2cb76dc7","Type":"ContainerStarted","Data":"a89ef75fe3c8781c4169410a1e4a36a180b680f946e68067e68d4c30f0c30878"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.483312 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" event={"ID":"52967242-12a0-48ed-a278-791b2cb76dc7","Type":"ContainerStarted","Data":"aac9f14f67618216a57568846d813f5d35933140aac83ca4f462eba3cf993c20"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.484400 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.485667 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6e236e2c-f743-4aac-8b58-f9dc13769abb-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-vxwd9\" (UID: \"6e236e2c-f743-4aac-8b58-f9dc13769abb\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.486380 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" event={"ID":"cf9c54fa-0743-40ed-85b7-ae38607f7265","Type":"ContainerStarted","Data":"75cbe7e5c43dfbcca2dba4494942383466d8d7a8ca697f76ced6eab437575552"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.491333 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.493112 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" event={"ID":"05d038d6-7946-445b-a1d3-c2d6f0044b7b","Type":"ContainerStarted","Data":"a276cffd2ee63e6617c0be8fb1808f5f8ffc1a22445eca80ac5f7f569f45bbf8"} Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.496499 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.502195 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8jt6\" (UniqueName: \"kubernetes.io/projected/98524d44-741a-4550-b185-5116a1498602-kube-api-access-v8jt6\") pod \"console-operator-58897d9998-ht8kr\" (UID: \"98524d44-741a-4550-b185-5116a1498602\") " pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.514500 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.520883 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4szsc\" (UniqueName: \"kubernetes.io/projected/04b5a53f-52f9-48e3-aa29-daaeeb41b76b-kube-api-access-4szsc\") pod \"service-ca-operator-777779d784-twxxw\" (UID: \"04b5a53f-52f9-48e3-aa29-daaeeb41b76b\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.527970 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.538301 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.550472 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxldr\" (UniqueName: \"kubernetes.io/projected/339c3294-8719-4942-9998-67605f596945-kube-api-access-qxldr\") pod \"dns-operator-744455d44c-5rg5n\" (UID: \"339c3294-8719-4942-9998-67605f596945\") " pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.556249 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.565685 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.574947 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.585735 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.587415 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.589316 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.605797 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.608834 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.610134 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-m9sb5"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.618141 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.626404 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.629991 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.641881 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6q4j\" (UniqueName: \"kubernetes.io/projected/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-kube-api-access-w6q4j\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.648340 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.662706 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-machine-approver-tls\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.667418 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.669107 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.673093 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.685818 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.706692 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.726009 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 21 00:08:29 crc kubenswrapper[4730]: W0221 00:08:29.728711 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod07a4c1db_3aad_4085_9eed_cd05bb34d6f2.slice/crio-268e397b93e3f7d4bf1bbedf0186c7dd85bf2448c45fb844b19f698811d9419c WatchSource:0}: Error finding container 268e397b93e3f7d4bf1bbedf0186c7dd85bf2448c45fb844b19f698811d9419c: Status 404 returned error can't find the container with id 268e397b93e3f7d4bf1bbedf0186c7dd85bf2448c45fb844b19f698811d9419c Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.738406 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-encryption-config\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.746994 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.755243 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6f272e98-1b2b-4c55-b095-5f0b2a2f8972-audit-policies\") pod \"apiserver-7bbb656c7d-zwmzp\" (UID: \"6f272e98-1b2b-4c55-b095-5f0b2a2f8972\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.758183 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.764467 4730 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.764544 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config podName:6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.764526772 +0000 UTC m=+163.055362502 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config") pod "machine-approver-56656f9798-5gb7h" (UID: "6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.764470 4730 configmap.go:193] Couldn't get configMap openshift-cluster-machine-approver/machine-approver-config: failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.764690 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config podName:6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1 nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.764665746 +0000 UTC m=+163.055501476 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config") pod "machine-approver-56656f9798-5gb7h" (UID: "6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1") : failed to sync configmap cache: timed out waiting for the condition Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.768167 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.790542 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.801476 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f98bd\" (UniqueName: \"kubernetes.io/projected/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-kube-api-access-f98bd\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.805385 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-mk5nk"] Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.807216 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.883790 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8785m\" (UniqueName: \"kubernetes.io/projected/4b9b1960-b0b9-4d2c-8227-52550eb224e6-kube-api-access-8785m\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.883994 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpzgq\" (UniqueName: \"kubernetes.io/projected/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-kube-api-access-hpzgq\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884062 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fbb922a-e91e-446d-aa76-9f68b8c141fd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884098 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0088317-03f9-47c6-8c33-b630f0299c21-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884135 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-brjr8\" (UniqueName: \"kubernetes.io/projected/031916fc-4c68-4c3f-8f77-a95d0a59b39d-kube-api-access-brjr8\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884179 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0088317-03f9-47c6-8c33-b630f0299c21-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884196 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpff9\" (UniqueName: \"kubernetes.io/projected/88933a34-ee44-455b-8ad4-40b3af8a3297-kube-api-access-lpff9\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884221 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88933a34-ee44-455b-8ad4-40b3af8a3297-available-featuregates\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884265 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d65lt\" (UniqueName: \"kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884282 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884306 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8nxf\" (UniqueName: \"kubernetes.io/projected/80742fc0-505d-400a-9261-90f3e5a23183-kube-api-access-x8nxf\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884330 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-config\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884361 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884376 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884400 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884415 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/80742fc0-505d-400a-9261-90f3e5a23183-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884451 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884476 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m7pd\" (UniqueName: \"kubernetes.io/projected/3bfc686e-e4f5-4be1-9014-1b37f03fc786-kube-api-access-5m7pd\") pod \"migrator-59844c95c7-fb9qr\" (UID: \"3bfc686e-e4f5-4be1-9014-1b37f03fc786\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884542 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/031916fc-4c68-4c3f-8f77-a95d0a59b39d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884559 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld7h5\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884575 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884625 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884640 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-images\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884654 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884720 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88933a34-ee44-455b-8ad4-40b3af8a3297-serving-cert\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884745 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884764 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-srv-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884816 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0088317-03f9-47c6-8c33-b630f0299c21-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884851 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884934 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884951 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjxkl\" (UniqueName: \"kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.884965 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwv5c\" (UniqueName: \"kubernetes.io/projected/0fbf3e59-6719-4583-a0aa-4badc34ecf3b-kube-api-access-hwv5c\") pod \"downloads-7954f5f757-bk7rn\" (UID: \"0fbf3e59-6719-4583-a0aa-4badc34ecf3b\") " pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.885220 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.885253 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fbb922a-e91e-446d-aa76-9f68b8c141fd-config\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.892893 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fbb922a-e91e-446d-aa76-9f68b8c141fd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.893101 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.903389 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.403370889 +0000 UTC m=+162.694206619 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.950729 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.996416 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:29 crc kubenswrapper[4730]: E0221 00:08:29.996592 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.496563622 +0000 UTC m=+162.787399352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.996938 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjxkl\" (UniqueName: \"kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.996969 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwv5c\" (UniqueName: \"kubernetes.io/projected/0fbf3e59-6719-4583-a0aa-4badc34ecf3b-kube-api-access-hwv5c\") pod \"downloads-7954f5f757-bk7rn\" (UID: \"0fbf3e59-6719-4583-a0aa-4badc34ecf3b\") " pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997014 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997062 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997098 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997131 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fbb922a-e91e-446d-aa76-9f68b8c141fd-config\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997156 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997201 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fbb922a-e91e-446d-aa76-9f68b8c141fd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997221 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-socket-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997287 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997352 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8785m\" (UniqueName: \"kubernetes.io/projected/4b9b1960-b0b9-4d2c-8227-52550eb224e6-kube-api-access-8785m\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997377 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpzgq\" (UniqueName: \"kubernetes.io/projected/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-kube-api-access-hpzgq\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997416 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-registration-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997578 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fbb922a-e91e-446d-aa76-9f68b8c141fd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997675 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0088317-03f9-47c6-8c33-b630f0299c21-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997715 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-plugins-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997738 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997756 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2s6v\" (UniqueName: \"kubernetes.io/projected/a03ce477-4d78-4f83-a197-f195cb6a0f6f-kube-api-access-t2s6v\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997777 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-brjr8\" (UniqueName: \"kubernetes.io/projected/031916fc-4c68-4c3f-8f77-a95d0a59b39d-kube-api-access-brjr8\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997795 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0088317-03f9-47c6-8c33-b630f0299c21-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997812 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpff9\" (UniqueName: \"kubernetes.io/projected/88933a34-ee44-455b-8ad4-40b3af8a3297-kube-api-access-lpff9\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997892 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88933a34-ee44-455b-8ad4-40b3af8a3297-available-featuregates\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997963 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-webhook-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.997992 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qghwl\" (UniqueName: \"kubernetes.io/projected/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-kube-api-access-qghwl\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998022 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-csi-data-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998078 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d65lt\" (UniqueName: \"kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998106 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9hx5\" (UniqueName: \"kubernetes.io/projected/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-kube-api-access-f9hx5\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998138 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-config-volume\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998205 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998249 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8nxf\" (UniqueName: \"kubernetes.io/projected/80742fc0-505d-400a-9261-90f3e5a23183-kube-api-access-x8nxf\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998275 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-config\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998316 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998340 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998358 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/88933a34-ee44-455b-8ad4-40b3af8a3297-available-featuregates\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998365 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998433 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/80742fc0-505d-400a-9261-90f3e5a23183-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998472 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-apiservice-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998502 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a03ce477-4d78-4f83-a197-f195cb6a0f6f-tmpfs\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998520 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998540 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-metrics-tls\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998557 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-mountpoint-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998595 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m7pd\" (UniqueName: \"kubernetes.io/projected/3bfc686e-e4f5-4be1-9014-1b37f03fc786-kube-api-access-5m7pd\") pod \"migrator-59844c95c7-fb9qr\" (UID: \"3bfc686e-e4f5-4be1-9014-1b37f03fc786\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998618 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-certs\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998658 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-289dt\" (UniqueName: \"kubernetes.io/projected/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-kube-api-access-289dt\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998704 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998710 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/031916fc-4c68-4c3f-8f77-a95d0a59b39d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998760 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld7h5\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998783 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998804 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxd87\" (UniqueName: \"kubernetes.io/projected/2baa63ee-4f71-41e4-830f-9589b69e1faa-kube-api-access-cxd87\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998826 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-node-bootstrap-token\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998844 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8z4f\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-kube-api-access-q8z4f\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998873 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998891 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-images\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.998973 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999057 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88933a34-ee44-455b-8ad4-40b3af8a3297-serving-cert\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999128 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999172 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-srv-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999196 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-cert\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999255 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0088317-03f9-47c6-8c33-b630f0299c21-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:29 crc kubenswrapper[4730]: I0221 00:08:29.999328 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.000575 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8fbb922a-e91e-446d-aa76-9f68b8c141fd-config\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.000807 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.500792145 +0000 UTC m=+162.791628105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.001033 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b0088317-03f9-47c6-8c33-b630f0299c21-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.004623 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-config\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.005651 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.006088 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.006946 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.006962 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.007785 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/031916fc-4c68-4c3f-8f77-a95d0a59b39d-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.007951 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/80742fc0-505d-400a-9261-90f3e5a23183-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.008410 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/031916fc-4c68-4c3f-8f77-a95d0a59b39d-images\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.008430 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.010128 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8fbb922a-e91e-446d-aa76-9f68b8c141fd-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.010358 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b0088317-03f9-47c6-8c33-b630f0299c21-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.010711 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-profile-collector-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.011104 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/4b9b1960-b0b9-4d2c-8227-52550eb224e6-srv-cert\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.011478 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.011766 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.012335 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/88933a34-ee44-455b-8ad4-40b3af8a3297-serving-cert\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.013245 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.015303 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.039053 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpff9\" (UniqueName: \"kubernetes.io/projected/88933a34-ee44-455b-8ad4-40b3af8a3297-kube-api-access-lpff9\") pod \"openshift-config-operator-7777fb866f-444fb\" (UID: \"88933a34-ee44-455b-8ad4-40b3af8a3297\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.075570 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-brjr8\" (UniqueName: \"kubernetes.io/projected/031916fc-4c68-4c3f-8f77-a95d0a59b39d-kube-api-access-brjr8\") pod \"machine-api-operator-5694c8668f-rmq4n\" (UID: \"031916fc-4c68-4c3f-8f77-a95d0a59b39d\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.090329 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwv5c\" (UniqueName: \"kubernetes.io/projected/0fbf3e59-6719-4583-a0aa-4badc34ecf3b-kube-api-access-hwv5c\") pod \"downloads-7954f5f757-bk7rn\" (UID: \"0fbf3e59-6719-4583-a0aa-4badc34ecf3b\") " pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.099952 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.100456 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.600439437 +0000 UTC m=+162.891275167 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100559 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxd87\" (UniqueName: \"kubernetes.io/projected/2baa63ee-4f71-41e4-830f-9589b69e1faa-kube-api-access-cxd87\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100586 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8z4f\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-kube-api-access-q8z4f\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100604 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-node-bootstrap-token\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100634 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100653 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-cert\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100688 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100706 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100735 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-socket-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.101249 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.60123776 +0000 UTC m=+162.892073490 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.102741 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-socket-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.103182 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-registration-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.103765 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b0088317-03f9-47c6-8c33-b630f0299c21-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-bhtnr\" (UID: \"b0088317-03f9-47c6-8c33-b630f0299c21\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.103856 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.100780 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-registration-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.108874 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-plugins-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.108937 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.108976 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2s6v\" (UniqueName: \"kubernetes.io/projected/a03ce477-4d78-4f83-a197-f195cb6a0f6f-kube-api-access-t2s6v\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.109008 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-webhook-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.109087 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qghwl\" (UniqueName: \"kubernetes.io/projected/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-kube-api-access-qghwl\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.109107 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-csi-data-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110077 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9hx5\" (UniqueName: \"kubernetes.io/projected/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-kube-api-access-f9hx5\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110120 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-config-volume\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110298 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-apiservice-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110323 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a03ce477-4d78-4f83-a197-f195cb6a0f6f-tmpfs\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110342 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-metrics-tls\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110360 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-mountpoint-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110387 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-certs\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110415 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-289dt\" (UniqueName: \"kubernetes.io/projected/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-kube-api-access-289dt\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.110672 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-plugins-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.111405 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-csi-data-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.112070 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-config-volume\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.112435 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/2baa63ee-4f71-41e4-830f-9589b69e1faa-mountpoint-dir\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.112827 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/a03ce477-4d78-4f83-a197-f195cb6a0f6f-tmpfs\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.113377 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.113619 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-node-bootstrap-token\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.118113 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-certs\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.118407 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-apiservice-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.121092 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-metrics-tls\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.122234 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-cert\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.131851 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a03ce477-4d78-4f83-a197-f195cb6a0f6f-webhook-cert\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.134097 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjxkl\" (UniqueName: \"kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl\") pod \"marketplace-operator-79b997595-vbnbc\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.149489 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpzgq\" (UniqueName: \"kubernetes.io/projected/ccbf3930-cbc6-4062-9d37-f6cb9b865da4-kube-api-access-hpzgq\") pod \"package-server-manager-789f6589d5-m5k9h\" (UID: \"ccbf3930-cbc6-4062-9d37-f6cb9b865da4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.170860 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8785m\" (UniqueName: \"kubernetes.io/projected/4b9b1960-b0b9-4d2c-8227-52550eb224e6-kube-api-access-8785m\") pod \"catalog-operator-68c6474976-qnt7r\" (UID: \"4b9b1960-b0b9-4d2c-8227-52550eb224e6\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.190368 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.190509 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d65lt\" (UniqueName: \"kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt\") pod \"collect-profiles-29527200-frqmr\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.205593 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.211821 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.212018 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.711971376 +0000 UTC m=+163.002807106 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.212199 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.212462 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.71245131 +0000 UTC m=+163.003287040 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.222637 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8fbb922a-e91e-446d-aa76-9f68b8c141fd-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-ghzgt\" (UID: \"8fbb922a-e91e-446d-aa76-9f68b8c141fd\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.251306 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m7pd\" (UniqueName: \"kubernetes.io/projected/3bfc686e-e4f5-4be1-9014-1b37f03fc786-kube-api-access-5m7pd\") pod \"migrator-59844c95c7-fb9qr\" (UID: \"3bfc686e-e4f5-4be1-9014-1b37f03fc786\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.253194 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.264810 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld7h5\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.267702 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.290110 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.292433 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8nxf\" (UniqueName: \"kubernetes.io/projected/80742fc0-505d-400a-9261-90f3e5a23183-kube-api-access-x8nxf\") pod \"control-plane-machine-set-operator-78cbb6b69f-zksgf\" (UID: \"80742fc0-505d-400a-9261-90f3e5a23183\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.293612 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.299752 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.313638 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.314570 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.814548313 +0000 UTC m=+163.105384043 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.324615 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8z4f\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-kube-api-access-q8z4f\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.356400 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.358860 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-m9mhf"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.359500 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxd87\" (UniqueName: \"kubernetes.io/projected/2baa63ee-4f71-41e4-830f-9589b69e1faa-kube-api-access-cxd87\") pod \"csi-hostpathplugin-n66hq\" (UID: \"2baa63ee-4f71-41e4-830f-9589b69e1faa\") " pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.365022 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.365182 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-289dt\" (UniqueName: \"kubernetes.io/projected/f6f2ad17-df9b-4c85-b092-0f21ecdd477f-kube-api-access-289dt\") pod \"ingress-canary-lpmdh\" (UID: \"f6f2ad17-df9b-4c85-b092-0f21ecdd477f\") " pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.373462 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.373533 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29527200-hgrnz"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.374012 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.383794 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.395358 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.396831 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-5xrdf\" (UID: \"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.406239 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.412686 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2s6v\" (UniqueName: \"kubernetes.io/projected/a03ce477-4d78-4f83-a197-f195cb6a0f6f-kube-api-access-t2s6v\") pod \"packageserver-d55dfcdfc-z6zvk\" (UID: \"a03ce477-4d78-4f83-a197-f195cb6a0f6f\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.415106 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.415466 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.420690 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:30.920674924 +0000 UTC m=+163.211510654 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.424430 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.434947 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-jpj6l"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.435187 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.447810 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qghwl\" (UniqueName: \"kubernetes.io/projected/b92b844b-d5d5-47c3-9f7e-1ada6f084a7e-kube-api-access-qghwl\") pod \"dns-default-djs5z\" (UID: \"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e\") " pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.448733 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9hx5\" (UniqueName: \"kubernetes.io/projected/ea224e51-5a71-4b8a-9f54-1ffa97b75f0a-kube-api-access-f9hx5\") pod \"machine-config-server-2q9tl\" (UID: \"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a\") " pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.456720 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-lpmdh" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.483668 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-bsnm7"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.483779 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-twxxw"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.483850 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-f5csk"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.484084 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-ht8kr"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.484155 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.484212 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.495938 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.505691 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.517416 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.517749 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.01772656 +0000 UTC m=+163.308562290 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.517997 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.518437 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.01841972 +0000 UTC m=+163.309255450 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.531486 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-4wrmp" event={"ID":"0d89fc71-51e8-4595-b9d4-fc0cd8103017","Type":"ContainerStarted","Data":"4f2e831c4e3ff47d5cf932586b17ef7e12700daae57d66e3731ab972ce77ed5c"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.542227 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.542341 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-5rg5n"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.570318 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" event={"ID":"6f272e98-1b2b-4c55-b095-5f0b2a2f8972","Type":"ContainerStarted","Data":"6c2e6eb9ce0923cd70ebfb092049df86becdb2e6f10a203cd742b4725fbce96a"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.579346 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29527200-hgrnz" event={"ID":"05c8cb3c-a432-4eef-895d-6154a07f3b90","Type":"ContainerStarted","Data":"afc839afa37bcbd02461a8f1e448c6cb7df96cc8b8cb0fb4ad7f6c1d155a05ae"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.604914 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" event={"ID":"f1012895-d7ed-4a60-95fa-8521f1e57b3d","Type":"ContainerStarted","Data":"2d582e8b36270f63b88fa25df75c00a87953946e083e7970598863f85cadac3c"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.604953 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" event={"ID":"f1012895-d7ed-4a60-95fa-8521f1e57b3d","Type":"ContainerStarted","Data":"f9e26b2b799d047a4f69d0829edf700b5e7f7df5e80db5a70d57cbf4cc731ea5"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.604964 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" event={"ID":"f1012895-d7ed-4a60-95fa-8521f1e57b3d","Type":"ContainerStarted","Data":"e76d972af0c78717ea21aa505006ee65383338c779496d1fe5a4a107e53ef9b2"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.607884 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" event={"ID":"2d012ccd-e3fb-4dbb-b375-e70f9d803919","Type":"ContainerStarted","Data":"642e99e862cc834551db5c421fd0f14c023748c8168789b014afc79ef769cc01"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.612222 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" event={"ID":"07a4c1db-3aad-4085-9eed-cd05bb34d6f2","Type":"ContainerStarted","Data":"cb236a115b0421c352188614ae515132000a9bed85bc60b6882d4389578f9d97"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.612271 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" event={"ID":"07a4c1db-3aad-4085-9eed-cd05bb34d6f2","Type":"ContainerStarted","Data":"268e397b93e3f7d4bf1bbedf0186c7dd85bf2448c45fb844b19f698811d9419c"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.613841 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" event={"ID":"2c79c4da-0098-43d4-83c9-0e140f37c8fe","Type":"ContainerStarted","Data":"67204b4584738a6ee19a98f79996d66c99cb46dd5198670f293fd91cdf157233"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.613870 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" event={"ID":"2c79c4da-0098-43d4-83c9-0e140f37c8fe","Type":"ContainerStarted","Data":"259a82d545700f4f94886b5a5231da41ee49c1f743c075f7d2fa7e1ee37359a1"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.617342 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" event={"ID":"e921b9fb-4785-4303-acd0-f04a0b5c4da8","Type":"ContainerStarted","Data":"1412f2f619639f91552d07e8e78f693c1734ab616068b0681d549e4a7ce0c8e6"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.618337 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rmq4n"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.618750 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.620639 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.120621576 +0000 UTC m=+163.411457306 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.630472 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" event={"ID":"05d038d6-7946-445b-a1d3-c2d6f0044b7b","Type":"ContainerStarted","Data":"d01b632fe8885a8f3a8acff4b4e86fb9ba63afd560da58d7dfa6e230d571a3f7"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.630534 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" event={"ID":"05d038d6-7946-445b-a1d3-c2d6f0044b7b","Type":"ContainerStarted","Data":"8311c52873ad76506440407509b933718499ce584bcfb5c6fac9d33b6bfdb98e"} Feb 21 00:08:30 crc kubenswrapper[4730]: W0221 00:08:30.664464 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod031916fc_4c68_4c3f_8f77_a95d0a59b39d.slice/crio-37ebc27e46fcb8858ee92266c52c3b5787488af3ba195e80afb727f5fdc26363 WatchSource:0}: Error finding container 37ebc27e46fcb8858ee92266c52c3b5787488af3ba195e80afb727f5fdc26363: Status 404 returned error can't find the container with id 37ebc27e46fcb8858ee92266c52c3b5787488af3ba195e80afb727f5fdc26363 Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.664471 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" event={"ID":"a132f6d0-9acf-4dd2-a773-2ae6640dae8b","Type":"ContainerStarted","Data":"999ac271687845e82540166b4d4577d98a9fd2e930bf8f35130092858f02d8c4"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.667131 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" event={"ID":"f28f3f0c-4af6-4f64-baef-b2fdff773582","Type":"ContainerStarted","Data":"3b2900640eece59c8967fca5b1f74babc16b9e580f1b3af5b7a08c07ba43672c"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.667160 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" event={"ID":"f28f3f0c-4af6-4f64-baef-b2fdff773582","Type":"ContainerStarted","Data":"4dab8f7a76442656ac7b1f503570c334feb57da503e16b938582c73d9d24bd7d"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.693442 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mk5nk" event={"ID":"d34ad48a-e9a9-47dc-a4b2-16edb864a63a","Type":"ContainerStarted","Data":"4cee800ff8306d516e7a067ba3c828c9f8934d34e28aee4c3c0ace284932bb8b"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.693518 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-mk5nk" event={"ID":"d34ad48a-e9a9-47dc-a4b2-16edb864a63a","Type":"ContainerStarted","Data":"8011b9502617e4b0a5782ffba3969f18deb44820f7f8e7bbf5b602929674565b"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.712402 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" event={"ID":"34eac1cb-c6b6-441b-bb23-f0b95fea55f2","Type":"ContainerStarted","Data":"82db817c64b0bdaeedd01ac86789b6c9527612e0f6684270b6eeb633be0764f3"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.712631 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.712647 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" event={"ID":"34eac1cb-c6b6-441b-bb23-f0b95fea55f2","Type":"ContainerStarted","Data":"fb869a18f3933b366b85da212b04ca85e224d89e33b8d24ffea74618f1131ae6"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.714338 4730 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z6jtl container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.714401 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" podUID="34eac1cb-c6b6-441b-bb23-f0b95fea55f2" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.715339 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" event={"ID":"cf9c54fa-0743-40ed-85b7-ae38607f7265","Type":"ContainerStarted","Data":"5e5cc3d04d9aa6416b05d9eed939ef8824c602fd2b4709cdea7faa531f23753a"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.718305 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.720003 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.721338 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.221322289 +0000 UTC m=+163.512158019 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.743932 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-2q9tl" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.758280 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" event={"ID":"52967242-12a0-48ed-a278-791b2cb76dc7","Type":"ContainerStarted","Data":"48112f109d3f368c908211e39a764ec9020e69f7fc769ee7125dceadb6e4c4cb"} Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.759649 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-444fb"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.809184 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-bk7rn"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.821473 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.821925 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.822009 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.823760 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.323739301 +0000 UTC m=+163.614575031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.825305 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.826237 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1-auth-proxy-config\") pod \"machine-approver-56656f9798-5gb7h\" (UID: \"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:30 crc kubenswrapper[4730]: W0221 00:08:30.841809 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod88933a34_ee44_455b_8ad4_40b3af8a3297.slice/crio-54db918d18e76af8279045a765f8228d04a4809f300d2e080539a3ffac99930c WatchSource:0}: Error finding container 54db918d18e76af8279045a765f8228d04a4809f300d2e080539a3ffac99930c: Status 404 returned error can't find the container with id 54db918d18e76af8279045a765f8228d04a4809f300d2e080539a3ffac99930c Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.862484 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.915588 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr"] Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.923748 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:30 crc kubenswrapper[4730]: E0221 00:08:30.924510 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.424496205 +0000 UTC m=+163.715331935 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:30 crc kubenswrapper[4730]: I0221 00:08:30.925559 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.025348 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.025510 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.026067 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.526051613 +0000 UTC m=+163.816887343 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.034147 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971-metrics-certs\") pod \"network-metrics-daemon-krrw8\" (UID: \"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971\") " pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.127859 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.129764 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.62863036 +0000 UTC m=+163.919466090 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.157252 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-krrw8" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.221086 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" podStartSLOduration=143.221069631 podStartE2EDuration="2m23.221069631s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.167666081 +0000 UTC m=+163.458501811" watchObservedRunningTime="2026-02-21 00:08:31.221069631 +0000 UTC m=+163.511905361" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.222738 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.223691 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-m9sb5" podStartSLOduration=143.223683327 podStartE2EDuration="2m23.223683327s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.210200383 +0000 UTC m=+163.501036113" watchObservedRunningTime="2026-02-21 00:08:31.223683327 +0000 UTC m=+163.514519057" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.230253 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.230822 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.231271 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.731255189 +0000 UTC m=+164.022090919 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.311696 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.337202 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.337550 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.837539554 +0000 UTC m=+164.128375284 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.338102 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.346647 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:31 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:31 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:31 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.346709 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.354473 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.363355 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.364625 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.401479 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-dtdxb" podStartSLOduration=143.401459722 podStartE2EDuration="2m23.401459722s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.357343403 +0000 UTC m=+163.648179133" watchObservedRunningTime="2026-02-21 00:08:31.401459722 +0000 UTC m=+163.692295452" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.407200 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.411695 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-mk5nk" podStartSLOduration=144.411683961 podStartE2EDuration="2m24.411683961s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.40995631 +0000 UTC m=+163.700792040" watchObservedRunningTime="2026-02-21 00:08:31.411683961 +0000 UTC m=+163.702519691" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.413537 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-djs5z"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.437804 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.437893 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.937866266 +0000 UTC m=+164.228701996 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.438067 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.438381 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:31.93836911 +0000 UTC m=+164.229204840 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.486981 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vctw7" podStartSLOduration=144.48696138 podStartE2EDuration="2m24.48696138s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.482711776 +0000 UTC m=+163.773547506" watchObservedRunningTime="2026-02-21 00:08:31.48696138 +0000 UTC m=+163.777797110" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.487194 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.537697 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-8fj4m" podStartSLOduration=143.537683082 podStartE2EDuration="2m23.537683082s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.52425774 +0000 UTC m=+163.815093470" watchObservedRunningTime="2026-02-21 00:08:31.537683082 +0000 UTC m=+163.828518812" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.542261 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.542613 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.042594806 +0000 UTC m=+164.333430536 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.544467 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-lpmdh"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.616144 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.619112 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-n66hq"] Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.645771 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.657470 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.157453432 +0000 UTC m=+164.448289162 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.703232 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-t246w" podStartSLOduration=143.703206189 podStartE2EDuration="2m23.703206189s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.694515735 +0000 UTC m=+163.985351465" watchObservedRunningTime="2026-02-21 00:08:31.703206189 +0000 UTC m=+163.994041919" Feb 21 00:08:31 crc kubenswrapper[4730]: W0221 00:08:31.714170 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podccbf3930_cbc6_4062_9d37_f6cb9b865da4.slice/crio-851ef6d2de9253d544245113bc186d213b76490738c052406f0d3605fe0f2947 WatchSource:0}: Error finding container 851ef6d2de9253d544245113bc186d213b76490738c052406f0d3605fe0f2947: Status 404 returned error can't find the container with id 851ef6d2de9253d544245113bc186d213b76490738c052406f0d3605fe0f2947 Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.757864 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.758272 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.258257017 +0000 UTC m=+164.549092737 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.816627 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lpmdh" event={"ID":"f6f2ad17-df9b-4c85-b092-0f21ecdd477f","Type":"ContainerStarted","Data":"5a4e8c8f4889a7ee52cb6c49344e3a23b7b495d1f3a7198d952f6e1e0b42c2f1"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.822690 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" event={"ID":"e921b9fb-4785-4303-acd0-f04a0b5c4da8","Type":"ContainerStarted","Data":"4ee33e42a5229f23f2cb8da52b9d3f2c272b225bbe4ae598a57254d09fff406c"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.860631 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.861000 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.360986159 +0000 UTC m=+164.651821889 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.870828 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" event={"ID":"ad7385c1-1a71-45ee-9758-d633abe871f2","Type":"ContainerStarted","Data":"0093fdb1f4bb3759a4f6012da44fa0bbce94e2bd5cd1dd4f7b1629ce691b3e55"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.916304 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" event={"ID":"a03ce477-4d78-4f83-a197-f195cb6a0f6f","Type":"ContainerStarted","Data":"aec7e295df19523715b99dd3325751fc8f51a7dffa38c26d14f482b39df592ed"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.932548 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-952vl" podStartSLOduration=144.932529299 podStartE2EDuration="2m24.932529299s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:31.89833552 +0000 UTC m=+164.189171250" watchObservedRunningTime="2026-02-21 00:08:31.932529299 +0000 UTC m=+164.223365029" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.962678 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:31 crc kubenswrapper[4730]: E0221 00:08:31.963928 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.463893856 +0000 UTC m=+164.754729586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.968195 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29527200-hgrnz" event={"ID":"05c8cb3c-a432-4eef-895d-6154a07f3b90","Type":"ContainerStarted","Data":"1808902e42f592dbdd29c600d68892edf306e266e013adc83dd1355b27867da2"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.979503 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" event={"ID":"98524d44-741a-4550-b185-5116a1498602","Type":"ContainerStarted","Data":"74574a05d8b5bad6520caa0770a7ef412d6822178835464eafb9df6bdcea92bf"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.979550 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" event={"ID":"98524d44-741a-4550-b185-5116a1498602","Type":"ContainerStarted","Data":"f645f9c2912ece6deabbfecd18aee1ff2c075fbd18947d40fe7ebaeea1a3eb2b"} Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.979981 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.998666 4730 patch_prober.go:28] interesting pod/console-operator-58897d9998-ht8kr container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 21 00:08:31 crc kubenswrapper[4730]: I0221 00:08:31.998723 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" podUID="98524d44-741a-4550-b185-5116a1498602" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.002261 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" event={"ID":"49975767-e31f-4b1b-9fbb-0e63abb8cd47","Type":"ContainerStarted","Data":"067215bd9fedf4d4ed676fc0cffbe3447971b701ceee52fada3836db553a3f3c"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.002294 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" event={"ID":"49975767-e31f-4b1b-9fbb-0e63abb8cd47","Type":"ContainerStarted","Data":"bb9ecb867ba9b500d31ba98e0e8425c4eed4ca58810a1527a2875936e4ea19b3"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.019290 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" event={"ID":"6c73c5be-27dd-4e76-9ad9-9008fd2a884d","Type":"ContainerStarted","Data":"0dcb8984b8f1c2fa1671feabafb47b411cab0f37daaab4ef4c1a8093b9b58a53"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.032653 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" event={"ID":"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1","Type":"ContainerStarted","Data":"6b6d8d9ea43ed6f7443fcb330f679ef356cf3f3ea795e54f9256c0eb39afdc5c"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.035268 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" podStartSLOduration=145.035246011 podStartE2EDuration="2m25.035246011s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.030134521 +0000 UTC m=+164.320970251" watchObservedRunningTime="2026-02-21 00:08:32.035246011 +0000 UTC m=+164.326081741" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.041977 4730 generic.go:334] "Generic (PLEG): container finished" podID="2d012ccd-e3fb-4dbb-b375-e70f9d803919" containerID="d040ef57b217decc128af6d5bbcb0b959caa8fd7c3eecd44b9f16a23a7eb8faa" exitCode=0 Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.042916 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" event={"ID":"2d012ccd-e3fb-4dbb-b375-e70f9d803919","Type":"ContainerDied","Data":"d040ef57b217decc128af6d5bbcb0b959caa8fd7c3eecd44b9f16a23a7eb8faa"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.044248 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" event={"ID":"6e236e2c-f743-4aac-8b58-f9dc13769abb","Type":"ContainerStarted","Data":"f5243ddfebdb133326ead30b6408f5267b370008a0f4a1dcded315a219c0837f"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.055080 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" event={"ID":"80742fc0-505d-400a-9261-90f3e5a23183","Type":"ContainerStarted","Data":"7dcd45745401614b8c2cf52fc6c82d783d01d2e8fc1e11d626658c346f8b1b39"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.055347 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-lpf96" podStartSLOduration=144.055336478 podStartE2EDuration="2m24.055336478s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.053085352 +0000 UTC m=+164.343921082" watchObservedRunningTime="2026-02-21 00:08:32.055336478 +0000 UTC m=+164.346172208" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.063956 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" event={"ID":"88933a34-ee44-455b-8ad4-40b3af8a3297","Type":"ContainerStarted","Data":"54db918d18e76af8279045a765f8228d04a4809f300d2e080539a3ffac99930c"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.064500 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.065271 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.565254818 +0000 UTC m=+164.856090548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.077849 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-krrw8"] Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.109218 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" event={"ID":"031916fc-4c68-4c3f-8f77-a95d0a59b39d","Type":"ContainerStarted","Data":"90fd5234553344aee25db274087f402a297e95d7551559281ee43724cebe3e98"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.109258 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" event={"ID":"031916fc-4c68-4c3f-8f77-a95d0a59b39d","Type":"ContainerStarted","Data":"37ebc27e46fcb8858ee92266c52c3b5787488af3ba195e80afb727f5fdc26363"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.125281 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bk7rn" event={"ID":"0fbf3e59-6719-4583-a0aa-4badc34ecf3b","Type":"ContainerStarted","Data":"ff90cf84baee505d06432ce1b30418ec3f586e24cdda9594090c52ebef39005d"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.144831 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" event={"ID":"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93","Type":"ContainerStarted","Data":"7b45c7f1b2f43a5374378b96f1859ce5e82c9f63df096ecac36b57b0070cfde6"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.163191 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" event={"ID":"4b9b1960-b0b9-4d2c-8227-52550eb224e6","Type":"ContainerStarted","Data":"a6262183db03f5de098899854af9b9fd06742d8c13c72e3028ad52e0c8645c8f"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.165477 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.165616 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.665582289 +0000 UTC m=+164.956418039 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.165737 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.166926 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.666880427 +0000 UTC m=+164.957716157 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.236679 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-djs5z" event={"ID":"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e","Type":"ContainerStarted","Data":"616c4dd1c19635371cded91c62f955823ce6024b06650d57a8556d0f0931e8c9"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.237537 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-4wrmp" podStartSLOduration=144.237523421 podStartE2EDuration="2m24.237523421s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.22960159 +0000 UTC m=+164.520437320" watchObservedRunningTime="2026-02-21 00:08:32.237523421 +0000 UTC m=+164.528359171" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.239353 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" event={"ID":"8fbb922a-e91e-446d-aa76-9f68b8c141fd","Type":"ContainerStarted","Data":"4e78a75b70a45bb0d24d85f5287be486a255768e1abf730064806ee483c6c4dd"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.242670 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" event={"ID":"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16","Type":"ContainerStarted","Data":"bee43a9c12754f36fa4184c96207458425677ef28efa19cb1350de1745b64ab8"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.245286 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2q9tl" event={"ID":"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a","Type":"ContainerStarted","Data":"53f72c0fb9af98eb3d3f04f5a5773e94774bae528aa7c06fc89a44a0e2b44312"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.246331 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" event={"ID":"b0088317-03f9-47c6-8c33-b630f0299c21","Type":"ContainerStarted","Data":"73bfe2f0c1e14c43c7e712a8d13d92e45a1ab8cf02089df6b5a46c83b52ae52d"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.250331 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" event={"ID":"3bfc686e-e4f5-4be1-9014-1b37f03fc786","Type":"ContainerStarted","Data":"130beb556c2d58c39d89da47fcf20657c9bfc8d5e077addc3181a5ca0f3a8c93"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.266843 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.271694 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" event={"ID":"ccbf3930-cbc6-4062-9d37-f6cb9b865da4","Type":"ContainerStarted","Data":"851ef6d2de9253d544245113bc186d213b76490738c052406f0d3605fe0f2947"} Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.278312 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.778276362 +0000 UTC m=+165.069112102 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.278481 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.278887 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.778870459 +0000 UTC m=+165.069706189 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.310587 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" event={"ID":"0804eedb-db79-4012-85c6-f3e572f93179","Type":"ContainerStarted","Data":"f7f916ac8b6ec59ecd004b9d06b9f3bec6d4ec078cc8d5c63b17dff3848b8e2f"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.311857 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.313473 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" event={"ID":"2eb73c39-8598-4fdf-afe6-314a98db6331","Type":"ContainerStarted","Data":"ddc5e9853a69d68afd7dba2a3df1c6e399fb843335ae2cf170de2bf436a5582f"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.324002 4730 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qjdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.324066 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.327361 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" event={"ID":"2baa63ee-4f71-41e4-830f-9589b69e1faa","Type":"ContainerStarted","Data":"3212365edffbe23b1f3b57f8d74fafcf004b07011d5d0961b128c74b1a272ca3"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.341590 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:32 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:32 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:32 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.342035 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.347470 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" event={"ID":"04b5a53f-52f9-48e3-aa29-daaeeb41b76b","Type":"ContainerStarted","Data":"92a917bc2f716d556c9eb68d4c9ae83e8c676ea6442f9f1548e1471713897fca"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.347519 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" event={"ID":"04b5a53f-52f9-48e3-aa29-daaeeb41b76b","Type":"ContainerStarted","Data":"74cc4861b55687ea3a7b5911bf3190254825566708b4bd33c44767d252d564a2"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.379732 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" event={"ID":"339c3294-8719-4942-9998-67605f596945","Type":"ContainerStarted","Data":"23214b49f3df709c0ff4e7f46aa2bbf144aa21797392cb0d3297eee10a4daeb7"} Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.388019 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.388343 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.888314277 +0000 UTC m=+165.179150007 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.390884 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.392068 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" event={"ID":"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf","Type":"ContainerStarted","Data":"40aead9276bffd87d97236d10c5dcc3450f0f40eb98c4e11687509f65707b7cb"} Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.392233 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.892210501 +0000 UTC m=+165.183046231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.420497 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z6jtl" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.493077 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.493210 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:32.993188392 +0000 UTC m=+165.284024122 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.500041 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.500497 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.000482635 +0000 UTC m=+165.291318365 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.600236 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podStartSLOduration=144.600221939 podStartE2EDuration="2m24.600221939s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.568052089 +0000 UTC m=+164.858887829" watchObservedRunningTime="2026-02-21 00:08:32.600221939 +0000 UTC m=+164.891057669" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.601615 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" podStartSLOduration=144.60161007 podStartE2EDuration="2m24.60161007s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.59820544 +0000 UTC m=+164.889041170" watchObservedRunningTime="2026-02-21 00:08:32.60161007 +0000 UTC m=+164.892445790" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.606798 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.607329 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.107310916 +0000 UTC m=+165.398146646 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.710017 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.710355 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.210342357 +0000 UTC m=+165.501178087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.723009 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rt9q4" podStartSLOduration=145.722993006 podStartE2EDuration="2m25.722993006s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.721522953 +0000 UTC m=+165.012358693" watchObservedRunningTime="2026-02-21 00:08:32.722993006 +0000 UTC m=+165.013828736" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.808860 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" podStartSLOduration=145.808844335 podStartE2EDuration="2m25.808844335s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.759586305 +0000 UTC m=+165.050422035" watchObservedRunningTime="2026-02-21 00:08:32.808844335 +0000 UTC m=+165.099680065" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.811380 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.811480 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.311458821 +0000 UTC m=+165.602294561 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.815025 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.815368 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.315354725 +0000 UTC m=+165.606190455 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.844007 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" podStartSLOduration=145.843991672 podStartE2EDuration="2m25.843991672s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.836518203 +0000 UTC m=+165.127353933" watchObservedRunningTime="2026-02-21 00:08:32.843991672 +0000 UTC m=+165.134827402" Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.915996 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:32 crc kubenswrapper[4730]: E0221 00:08:32.916270 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.416255593 +0000 UTC m=+165.707091323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:32 crc kubenswrapper[4730]: I0221 00:08:32.921361 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-twxxw" podStartSLOduration=144.921344212 podStartE2EDuration="2m24.921344212s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.880088586 +0000 UTC m=+165.170924316" watchObservedRunningTime="2026-02-21 00:08:32.921344212 +0000 UTC m=+165.212179942" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.017892 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.018384 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.518373037 +0000 UTC m=+165.809208767 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.120986 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.121229 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.621199101 +0000 UTC m=+165.912034831 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.121435 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.122336 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.622320664 +0000 UTC m=+165.913156594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.224706 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.225435 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.725414246 +0000 UTC m=+166.016249976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.330249 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.330851 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.830835996 +0000 UTC m=+166.121671726 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.348631 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:33 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:33 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:33 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.348686 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.407137 4730 generic.go:334] "Generic (PLEG): container finished" podID="6f272e98-1b2b-4c55-b095-5f0b2a2f8972" containerID="2c252a0276577ce7aec88f3e6cbce30035fb6a2738a8bfffd997aad0cc09e089" exitCode=0 Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.407361 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" event={"ID":"6f272e98-1b2b-4c55-b095-5f0b2a2f8972","Type":"ContainerDied","Data":"2c252a0276577ce7aec88f3e6cbce30035fb6a2738a8bfffd997aad0cc09e089"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.416694 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-2q9tl" event={"ID":"ea224e51-5a71-4b8a-9f54-1ffa97b75f0a","Type":"ContainerStarted","Data":"46a4e80c1e183eb45f9b7ec40b887962c79199738ea8632ca0b4c15359e05ddb"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.424112 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" event={"ID":"6c73c5be-27dd-4e76-9ad9-9008fd2a884d","Type":"ContainerStarted","Data":"a65d9e17aac5d697a17704052c8a3d83a53e3fddc489a0732b6323966fcff9ac"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.432009 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.432462 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:33.932443565 +0000 UTC m=+166.223279295 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.444508 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29527200-hgrnz" podStartSLOduration=146.444488437 podStartE2EDuration="2m26.444488437s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:32.923242867 +0000 UTC m=+165.214078597" watchObservedRunningTime="2026-02-21 00:08:33.444488437 +0000 UTC m=+165.735324177" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.447034 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-djs5z" event={"ID":"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e","Type":"ContainerStarted","Data":"69003cd404939bea11a921862c9632c09d27c2464dc30bc86732517d0086ac34"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.462324 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-f5csk" event={"ID":"2eb73c39-8598-4fdf-afe6-314a98db6331","Type":"ContainerStarted","Data":"42ae5ad5a4872d19bdde7ed1da55a7dc320b3c7e12d69aff28a0d66825cf5fbb"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.482865 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-2q9tl" podStartSLOduration=7.482836367 podStartE2EDuration="7.482836367s" podCreationTimestamp="2026-02-21 00:08:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.479132019 +0000 UTC m=+165.769967749" watchObservedRunningTime="2026-02-21 00:08:33.482836367 +0000 UTC m=+165.773672097" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.494702 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" event={"ID":"3bfc686e-e4f5-4be1-9014-1b37f03fc786","Type":"ContainerStarted","Data":"435c095e51ad3960c3e9f101854b423a242a036763cbc0bbff85a823472a6d3d"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.506265 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" event={"ID":"ccbf3930-cbc6-4062-9d37-f6cb9b865da4","Type":"ContainerStarted","Data":"ad4bfe2af96409ef924edab281478831d7c3ea917229a868844b19a76576fa07"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.508199 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" event={"ID":"4b9b1960-b0b9-4d2c-8227-52550eb224e6","Type":"ContainerStarted","Data":"b3f7e99f976ab1131418ad98d22cf1e40633dd39ef0429ed62ccda9d7ff24e66"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.509190 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.527187 4730 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qnt7r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.527236 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" podUID="4b9b1960-b0b9-4d2c-8227-52550eb224e6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.534002 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.534339 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.034325572 +0000 UTC m=+166.325161302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.537171 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" event={"ID":"a03ce477-4d78-4f83-a197-f195cb6a0f6f","Type":"ContainerStarted","Data":"0f4e5f3ddd52f9a295eaa6231178f2adc132554329acb0cc7cebc576b4196003"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.538077 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.554958 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" podStartSLOduration=145.554943744 podStartE2EDuration="2m25.554943744s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.552512253 +0000 UTC m=+165.843347993" watchObservedRunningTime="2026-02-21 00:08:33.554943744 +0000 UTC m=+165.845779474" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.561053 4730 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-z6zvk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.561111 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" podUID="a03ce477-4d78-4f83-a197-f195cb6a0f6f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.572530 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" event={"ID":"6e025d1e-4c07-4cd2-8cb2-a34ae7641e16","Type":"ContainerStarted","Data":"ec65b9fbe86017d42a3a09ff646b799d82a3a05a4baf288768ce162f0f43d749"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.582339 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" event={"ID":"9ab92648-b7ad-412f-b9cd-4c8fb9f0fb93","Type":"ContainerStarted","Data":"9ad2abf80e2d85a734422c1f2fd5b2a1f5f7a5f7748393c6201c27588a20eecd"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.587944 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" event={"ID":"0804eedb-db79-4012-85c6-f3e572f93179","Type":"ContainerStarted","Data":"15bdb7ede88d5b9da751da95e920fa2ddb139ffba22a0255d6705009aacd8f09"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.588971 4730 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qjdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.588994 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.599979 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" event={"ID":"8fbb922a-e91e-446d-aa76-9f68b8c141fd","Type":"ContainerStarted","Data":"bd9fa85fa8d94e8470f1cc8e2ae814e45e7c5995e3dad8bcc30a841e66ec035d"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.600829 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" podStartSLOduration=145.600819275 podStartE2EDuration="2m25.600819275s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.600504186 +0000 UTC m=+165.891339936" watchObservedRunningTime="2026-02-21 00:08:33.600819275 +0000 UTC m=+165.891655005" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.626483 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-lpmdh" event={"ID":"f6f2ad17-df9b-4c85-b092-0f21ecdd477f","Type":"ContainerStarted","Data":"e7e6e7e52d6c75645d2d76c1e74d7f359eac50857be50d24b8b3ae4459529750"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.635834 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.636814 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.136795326 +0000 UTC m=+166.427631056 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.672204 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krrw8" event={"ID":"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971","Type":"ContainerStarted","Data":"c90e47ea1f9db85d151abbe8156807e333d0abde174188d6b638abc668efedf9"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.674651 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-5xrdf" podStartSLOduration=146.674621671 podStartE2EDuration="2m26.674621671s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.650602769 +0000 UTC m=+165.941438499" watchObservedRunningTime="2026-02-21 00:08:33.674621671 +0000 UTC m=+165.965457401" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.679093 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" event={"ID":"b0088317-03f9-47c6-8c33-b630f0299c21","Type":"ContainerStarted","Data":"4416ffe5b03c4543be63dda815ff800648b242747b7c845d33dc363a308d5fbf"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.681627 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-ghzgt" podStartSLOduration=145.681596045 podStartE2EDuration="2m25.681596045s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.67183961 +0000 UTC m=+165.962675360" watchObservedRunningTime="2026-02-21 00:08:33.681596045 +0000 UTC m=+165.972431775" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.688224 4730 generic.go:334] "Generic (PLEG): container finished" podID="88933a34-ee44-455b-8ad4-40b3af8a3297" containerID="fccb0b317cf3e1c5d2464b68cb06f44ae2bf99d6151075cb258a36d103eaecdd" exitCode=0 Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.688341 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" event={"ID":"88933a34-ee44-455b-8ad4-40b3af8a3297","Type":"ContainerDied","Data":"fccb0b317cf3e1c5d2464b68cb06f44ae2bf99d6151075cb258a36d103eaecdd"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.700522 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-bsnm7" podStartSLOduration=145.700486297 podStartE2EDuration="2m25.700486297s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.695054408 +0000 UTC m=+165.985890138" watchObservedRunningTime="2026-02-21 00:08:33.700486297 +0000 UTC m=+165.991322037" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.720786 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-bhtnr" podStartSLOduration=145.72077042 podStartE2EDuration="2m25.72077042s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.720320927 +0000 UTC m=+166.011156657" watchObservedRunningTime="2026-02-21 00:08:33.72077042 +0000 UTC m=+166.011606150" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.729412 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" event={"ID":"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf","Type":"ContainerStarted","Data":"3f694844b8366bd91f34a6a1fb0cc86b4182dd1145cac0b321e62387d5297478"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.730435 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.732302 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" event={"ID":"031916fc-4c68-4c3f-8f77-a95d0a59b39d","Type":"ContainerStarted","Data":"cc43c927bd7cbc446099ffd377b74a77455a3416ac3f3d4723dff46fdeb372a9"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.740321 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.743453 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.243438522 +0000 UTC m=+166.534274252 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.743559 4730 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lj2pw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.743819 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": dial tcp 10.217.0.12:8443: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.744513 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" event={"ID":"49975767-e31f-4b1b-9fbb-0e63abb8cd47","Type":"ContainerStarted","Data":"d5baba62f792ca9118d93f09d119da859adb442808b834477f42506eba136b02"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.749438 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" event={"ID":"6e236e2c-f743-4aac-8b58-f9dc13769abb","Type":"ContainerStarted","Data":"ebd0e5bbdd438f35719907c55b4ffa08a2215517bf806bd4e6a4624eb265d0bb"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.760143 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" event={"ID":"80742fc0-505d-400a-9261-90f3e5a23183","Type":"ContainerStarted","Data":"9a70f066dd06bebe9806cd164e22bb64b27b91c6fcfdfc43578745766c44f884"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.770650 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" event={"ID":"339c3294-8719-4942-9998-67605f596945","Type":"ContainerStarted","Data":"262aaf4d857795608b6af460c1692d7dc92a160e42adc2c18aeb406ab90289a3"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.773408 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-lpmdh" podStartSLOduration=7.773389107 podStartE2EDuration="7.773389107s" podCreationTimestamp="2026-02-21 00:08:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.769601137 +0000 UTC m=+166.060436867" watchObservedRunningTime="2026-02-21 00:08:33.773389107 +0000 UTC m=+166.064224847" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.784063 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" event={"ID":"ad7385c1-1a71-45ee-9758-d633abe871f2","Type":"ContainerStarted","Data":"8981e64d6fb05105d12163729c0ea0c4a49d13888c44bf8ee61467dde5bf1af7"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.794391 4730 patch_prober.go:28] interesting pod/console-operator-58897d9998-ht8kr container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.794443 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" podUID="98524d44-741a-4550-b185-5116a1498602" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.794934 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bk7rn" event={"ID":"0fbf3e59-6719-4583-a0aa-4badc34ecf3b","Type":"ContainerStarted","Data":"691135f2d846b50187897515760760f57da831222faa7ad2a02c133c5ab2c08d"} Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.795720 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.808788 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-rmq4n" podStartSLOduration=145.808775411 podStartE2EDuration="2m25.808775411s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.807799663 +0000 UTC m=+166.098635393" watchObservedRunningTime="2026-02-21 00:08:33.808775411 +0000 UTC m=+166.099611141" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.823792 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.823882 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.841462 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.844757 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.344739202 +0000 UTC m=+166.635574932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.854851 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-jpj6l" podStartSLOduration=145.854824997 podStartE2EDuration="2m25.854824997s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.85459144 +0000 UTC m=+166.145427170" watchObservedRunningTime="2026-02-21 00:08:33.854824997 +0000 UTC m=+166.145660727" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.910234 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-vxwd9" podStartSLOduration=145.910214105 podStartE2EDuration="2m25.910214105s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.909544346 +0000 UTC m=+166.200380096" watchObservedRunningTime="2026-02-21 00:08:33.910214105 +0000 UTC m=+166.201049835" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.939924 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-zksgf" podStartSLOduration=145.939909583 podStartE2EDuration="2m25.939909583s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.937119291 +0000 UTC m=+166.227955021" watchObservedRunningTime="2026-02-21 00:08:33.939909583 +0000 UTC m=+166.230745313" Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.950204 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:33 crc kubenswrapper[4730]: E0221 00:08:33.952507 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.452493521 +0000 UTC m=+166.743329251 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:33 crc kubenswrapper[4730]: I0221 00:08:33.967810 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" podStartSLOduration=145.967793258 podStartE2EDuration="2m25.967793258s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:33.967295523 +0000 UTC m=+166.258131253" watchObservedRunningTime="2026-02-21 00:08:33.967793258 +0000 UTC m=+166.258628978" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.002281 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" podStartSLOduration=146.002263275 podStartE2EDuration="2m26.002263275s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.001215134 +0000 UTC m=+166.292050864" watchObservedRunningTime="2026-02-21 00:08:34.002263275 +0000 UTC m=+166.293099015" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.065856 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.068594 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.568573202 +0000 UTC m=+166.859408932 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.112995 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.113464 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.613449654 +0000 UTC m=+166.904285394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.215320 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.215787 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.715769893 +0000 UTC m=+167.006605623 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.317100 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.317397 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.817382792 +0000 UTC m=+167.108218522 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.343782 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:34 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:34 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:34 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.344052 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.418169 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.418323 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.918304871 +0000 UTC m=+167.209140601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.418412 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.418693 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:34.918684662 +0000 UTC m=+167.209520392 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.519559 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.519718 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.019694784 +0000 UTC m=+167.310530524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.519812 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.520064 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.020056324 +0000 UTC m=+167.310892054 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.621364 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.621497 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.121478538 +0000 UTC m=+167.412314268 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.621636 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.621965 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.121957512 +0000 UTC m=+167.412793242 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.722510 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.722882 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.22286825 +0000 UTC m=+167.513703980 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.799780 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" event={"ID":"2d012ccd-e3fb-4dbb-b375-e70f9d803919","Type":"ContainerStarted","Data":"a4daee8b03d8473d15ad4294c07516a914396e15954c533423dc7ff924db3ab9"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.799820 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" event={"ID":"2d012ccd-e3fb-4dbb-b375-e70f9d803919","Type":"ContainerStarted","Data":"71a849ed1a960f851b072303d81c54cc3c3ff8f919eff524980388b82e0e3887"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.802094 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krrw8" event={"ID":"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971","Type":"ContainerStarted","Data":"1e80fcce0863cd2a6bebb16a9f6f96f15dcd254e545135c71aca21d263d8a782"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.802119 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-krrw8" event={"ID":"73bd3f3f-26f6-4db9-b4f7-8f0e2a8a1971","Type":"ContainerStarted","Data":"14c157d118e80442283fa2a9d1b2f1374160b7bb028f57c71e87ae6a31cd05d4"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.804882 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" event={"ID":"3bfc686e-e4f5-4be1-9014-1b37f03fc786","Type":"ContainerStarted","Data":"3af688249ddaa27ad37e9a6900fd08acb12a68e9b3f8e05a21815b5bafa5a7c2"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.806755 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" event={"ID":"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1","Type":"ContainerStarted","Data":"b269b188a3557b0b3434fddca7e03656e76a136409aa76deb5cddc945151b282"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.806781 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" event={"ID":"6df3c85f-e67b-4fb7-aa77-f5f9a9255aa1","Type":"ContainerStarted","Data":"1928d1cdc42dc9da89dc66f2673e69e3f63cbb6e7bfb34ec962c0fe96d44d5d6"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.809137 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-5rg5n" event={"ID":"339c3294-8719-4942-9998-67605f596945","Type":"ContainerStarted","Data":"1ccd1f5fa58b571a90259daa98eb5c7aacc66d88927d031904a19cd0e2ba36c7"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.810813 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-djs5z" event={"ID":"b92b844b-d5d5-47c3-9f7e-1ada6f084a7e","Type":"ContainerStarted","Data":"2f27f9d7a2563df635eeae78bf4031885bf3d702e88c6ab7cbe8ea23ff8f31cb"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.811148 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.812578 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" event={"ID":"ccbf3930-cbc6-4062-9d37-f6cb9b865da4","Type":"ContainerStarted","Data":"3f28bffa2e12e55760d2595ddac0a75c22bbcc6f1b66d389d2e7e07bc11095bd"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.812921 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.815191 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" event={"ID":"6f272e98-1b2b-4c55-b095-5f0b2a2f8972","Type":"ContainerStarted","Data":"920cf08457dff8cc6061e516415534f1e53893eee1ba7eb7fed172e5d7a234c5"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.817742 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" event={"ID":"88933a34-ee44-455b-8ad4-40b3af8a3297","Type":"ContainerStarted","Data":"c57f86b34fa4b06a5baf11e6c2a685d9d5ce59a2905e8fa77acd064a93007a6b"} Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818429 4730 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qjdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818475 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818478 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818533 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818587 4730 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-z6zvk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818609 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" podUID="a03ce477-4d78-4f83-a197-f195cb6a0f6f" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.37:5443/healthz\": dial tcp 10.217.0.37:5443: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818669 4730 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qnt7r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818684 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" podUID="4b9b1960-b0b9-4d2c-8227-52550eb224e6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818883 4730 patch_prober.go:28] interesting pod/console-operator-58897d9998-ht8kr container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.818921 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" podUID="98524d44-741a-4550-b185-5116a1498602" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.823996 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.825298 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.325270682 +0000 UTC m=+167.616106412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.878451 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-bk7rn" podStartSLOduration=147.878434336 podStartE2EDuration="2m27.878434336s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.068358886 +0000 UTC m=+166.359194616" watchObservedRunningTime="2026-02-21 00:08:34.878434336 +0000 UTC m=+167.169270056" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.880229 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" podStartSLOduration=147.880221978 podStartE2EDuration="2m27.880221978s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.875773828 +0000 UTC m=+167.166609558" watchObservedRunningTime="2026-02-21 00:08:34.880221978 +0000 UTC m=+167.171057708" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.895377 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-5gb7h" podStartSLOduration=147.895366651 podStartE2EDuration="2m27.895366651s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.892725374 +0000 UTC m=+167.183561104" watchObservedRunningTime="2026-02-21 00:08:34.895366651 +0000 UTC m=+167.186202381" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.905917 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-fb9qr" podStartSLOduration=146.905894188 podStartE2EDuration="2m26.905894188s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.904398505 +0000 UTC m=+167.195234235" watchObservedRunningTime="2026-02-21 00:08:34.905894188 +0000 UTC m=+167.196729918" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.924942 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" podStartSLOduration=146.924893223 podStartE2EDuration="2m26.924893223s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.923643447 +0000 UTC m=+167.214479187" watchObservedRunningTime="2026-02-21 00:08:34.924893223 +0000 UTC m=+167.215728953" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.924985 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.925141 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.42511269 +0000 UTC m=+167.715948420 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.925447 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:34 crc kubenswrapper[4730]: E0221 00:08:34.931976 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.43195679 +0000 UTC m=+167.722792520 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.951093 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.951174 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.953376 4730 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-zwmzp container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.953614 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" podUID="6f272e98-1b2b-4c55-b095-5f0b2a2f8972" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.5:8443/livez\": dial tcp 10.217.0.5:8443: connect: connection refused" Feb 21 00:08:34 crc kubenswrapper[4730]: I0221 00:08:34.980352 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-krrw8" podStartSLOduration=147.980336293 podStartE2EDuration="2m27.980336293s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:34.97919947 +0000 UTC m=+167.270035200" watchObservedRunningTime="2026-02-21 00:08:34.980336293 +0000 UTC m=+167.271172023" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.006825 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-djs5z" podStartSLOduration=8.006810537 podStartE2EDuration="8.006810537s" podCreationTimestamp="2026-02-21 00:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:35.004657534 +0000 UTC m=+167.295493264" watchObservedRunningTime="2026-02-21 00:08:35.006810537 +0000 UTC m=+167.297646267" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.019528 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" podStartSLOduration=147.019512618 podStartE2EDuration="2m27.019512618s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:35.018860809 +0000 UTC m=+167.309696529" watchObservedRunningTime="2026-02-21 00:08:35.019512618 +0000 UTC m=+167.310348348" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.024459 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.027214 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.027354 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.527326386 +0000 UTC m=+167.818162116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.027562 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.027823 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.527810741 +0000 UTC m=+167.818646471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.035264 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" podStartSLOduration=147.035247748 podStartE2EDuration="2m27.035247748s" podCreationTimestamp="2026-02-21 00:06:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:35.032784806 +0000 UTC m=+167.323620536" watchObservedRunningTime="2026-02-21 00:08:35.035247748 +0000 UTC m=+167.326083478" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.054243 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" podStartSLOduration=148.054224632 podStartE2EDuration="2m28.054224632s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:35.052380039 +0000 UTC m=+167.343215789" watchObservedRunningTime="2026-02-21 00:08:35.054224632 +0000 UTC m=+167.345060352" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.128677 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.128884 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.628851603 +0000 UTC m=+167.919687333 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.129193 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.129576 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.629568094 +0000 UTC m=+167.920403824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.229712 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.230006 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.729992888 +0000 UTC m=+168.020828608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.331126 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.331483 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.831466213 +0000 UTC m=+168.122302153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.343955 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:35 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:35 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:35 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.343999 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.432421 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.432654 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.932626919 +0000 UTC m=+168.223462639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.433086 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.433416 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:35.933406692 +0000 UTC m=+168.224242422 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.534424 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.534859 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.034843046 +0000 UTC m=+168.325678776 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.636291 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.636912 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.136856817 +0000 UTC m=+168.427692727 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.737993 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.738206 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.238179267 +0000 UTC m=+168.529014997 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.738240 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.738565 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.238556248 +0000 UTC m=+168.529391978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.775269 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.775328 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.839747 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.840089 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.340040763 +0000 UTC m=+168.630876493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.840228 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.840681 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.340669212 +0000 UTC m=+168.631504992 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.870611 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" event={"ID":"2baa63ee-4f71-41e4-830f-9589b69e1faa","Type":"ContainerStarted","Data":"22974fcdb8c7197df42c616b949b03c5294d2b5d86c9a22a3e60f435898fcdb4"} Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.874383 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.875171 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.875234 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.876022 4730 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-qnt7r container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" start-of-body= Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.876140 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" podUID="4b9b1960-b0b9-4d2c-8227-52550eb224e6" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/healthz\": dial tcp 10.217.0.35:8443: connect: connection refused" Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.941858 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.942085 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.442057424 +0000 UTC m=+168.732893164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:35 crc kubenswrapper[4730]: I0221 00:08:35.942593 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:35 crc kubenswrapper[4730]: E0221 00:08:35.944072 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.444051583 +0000 UTC m=+168.734887543 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.044426 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.044820 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.544768136 +0000 UTC m=+168.835603866 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.146762 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.147416 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.647375464 +0000 UTC m=+168.938211204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.247827 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.248286 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.748262052 +0000 UTC m=+169.039097782 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.340760 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:36 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:36 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:36 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.340848 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.349989 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.350421 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.850402946 +0000 UTC m=+169.141238676 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.451479 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.451707 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.951676255 +0000 UTC m=+169.242511985 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.452163 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.452597 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:36.952578032 +0000 UTC m=+169.243413762 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.553470 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.554500 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.054466369 +0000 UTC m=+169.345302099 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.639360 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-z6zvk" Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.655010 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.655601 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.155572433 +0000 UTC m=+169.446408353 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.756640 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.756931 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.256861782 +0000 UTC m=+169.547697512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.757291 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.757751 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.257742537 +0000 UTC m=+169.548578267 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.858606 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.858782 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.358757489 +0000 UTC m=+169.649593219 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.858881 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.859233 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.359225443 +0000 UTC m=+169.650061173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.960246 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.960425 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.460394529 +0000 UTC m=+169.751230259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:36 crc kubenswrapper[4730]: I0221 00:08:36.960693 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:36 crc kubenswrapper[4730]: E0221 00:08:36.961527 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.461499691 +0000 UTC m=+169.752335421 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.061689 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.061822 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.561803062 +0000 UTC m=+169.852638792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.061879 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.062147 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.562139862 +0000 UTC m=+169.852975592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.162996 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.163477 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.663457182 +0000 UTC m=+169.954292912 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.264979 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.265375 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.7653612 +0000 UTC m=+170.056196930 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.302367 4730 csr.go:261] certificate signing request csr-f65c8 is approved, waiting to be issued Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.311420 4730 csr.go:257] certificate signing request csr-f65c8 is issued Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.342754 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:37 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:37 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:37 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.342831 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.366429 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.366776 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.866756662 +0000 UTC m=+170.157592392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.366893 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.367187 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.867175495 +0000 UTC m=+170.158011225 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.468095 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.468986 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:37.968954458 +0000 UTC m=+170.259790218 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.570615 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.571184 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.071163755 +0000 UTC m=+170.361999485 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.671658 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.672054 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.172012932 +0000 UTC m=+170.462848662 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.773018 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.773665 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.273634241 +0000 UTC m=+170.564470181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.874676 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.874941 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.37489755 +0000 UTC m=+170.665733280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.874973 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.875287 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.375275111 +0000 UTC m=+170.666110841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.880109 4730 generic.go:334] "Generic (PLEG): container finished" podID="ad7385c1-1a71-45ee-9758-d633abe871f2" containerID="8981e64d6fb05105d12163729c0ea0c4a49d13888c44bf8ee61467dde5bf1af7" exitCode=0 Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.880150 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" event={"ID":"ad7385c1-1a71-45ee-9758-d633abe871f2","Type":"ContainerDied","Data":"8981e64d6fb05105d12163729c0ea0c4a49d13888c44bf8ee61467dde5bf1af7"} Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.914095 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.915282 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.917219 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.930586 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.976073 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.976275 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.476246591 +0000 UTC m=+170.767082311 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.976456 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.976507 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.976605 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drr9w\" (UniqueName: \"kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:37 crc kubenswrapper[4730]: I0221 00:08:37.976702 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:37 crc kubenswrapper[4730]: E0221 00:08:37.976783 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.476773967 +0000 UTC m=+170.767609697 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078229 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.078378 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.578352595 +0000 UTC m=+170.869188325 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078447 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078511 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078532 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078587 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drr9w\" (UniqueName: \"kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.078766 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.578757826 +0000 UTC m=+170.869593556 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.078877 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.079298 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.103509 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drr9w\" (UniqueName: \"kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w\") pod \"community-operators-d4xfg\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.114617 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.116181 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.118925 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.126452 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.179861 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.180021 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.679995105 +0000 UTC m=+170.970830835 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.180165 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.180187 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.180309 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vtzm\" (UniqueName: \"kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.180593 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.180971 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.680959433 +0000 UTC m=+170.971795163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.227293 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281195 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.281330 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.781309945 +0000 UTC m=+171.072145675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281405 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281423 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281439 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vtzm\" (UniqueName: \"kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281499 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.281742 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.781734347 +0000 UTC m=+171.072570077 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281871 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.281963 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.307331 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.308935 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.310962 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vtzm\" (UniqueName: \"kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm\") pod \"certified-operators-rk2d5\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.313201 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-21 00:03:37 +0000 UTC, rotation deadline is 2026-12-20 05:18:04.529709453 +0000 UTC Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.313237 4730 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7253h9m26.216474475s for next certificate rotation Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.341186 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.342230 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:38 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:38 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:38 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.342283 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.382938 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.383133 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.883080599 +0000 UTC m=+171.173916389 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.383285 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l9qmc\" (UniqueName: \"kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.383480 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.384334 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.384398 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.384830 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.884817759 +0000 UTC m=+171.175653489 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.444576 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.485269 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.485475 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.98544459 +0000 UTC m=+171.276280330 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.485506 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.485536 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.485594 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l9qmc\" (UniqueName: \"kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.485635 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.485845 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:38.985835281 +0000 UTC m=+171.276671031 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.486307 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.488802 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.504334 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.505392 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.512362 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.517543 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l9qmc\" (UniqueName: \"kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc\") pod \"community-operators-pjxx6\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.573628 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.574364 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.575842 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.576014 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.586203 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.587623 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.588129 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.588173 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qlhc\" (UniqueName: \"kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.588252 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.588352 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.088340016 +0000 UTC m=+171.379175746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.602321 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.634514 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:08:38 crc kubenswrapper[4730]: W0221 00:08:38.635668 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod17855419_b9b5_4c88_a004_694eb8320ea7.slice/crio-3972471d1a40ed3de9c1598bafd20ed04872fc57aa083acb76e6889c2d562eb5 WatchSource:0}: Error finding container 3972471d1a40ed3de9c1598bafd20ed04872fc57aa083acb76e6889c2d562eb5: Status 404 returned error can't find the container with id 3972471d1a40ed3de9c1598bafd20ed04872fc57aa083acb76e6889c2d562eb5 Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.688943 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.688988 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.689046 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.689112 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.689463 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.689662 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.189645696 +0000 UTC m=+171.480481416 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.689653 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.689806 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qlhc\" (UniqueName: \"kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.690016 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.712896 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:08:38 crc kubenswrapper[4730]: W0221 00:08:38.723169 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8b1b254_42f2_4ff9_8971_13528ab4f8d0.slice/crio-79f0b515be9b1050e7f9e9c36c2d89d8a44277033f208a47aa8716e3a19a828d WatchSource:0}: Error finding container 79f0b515be9b1050e7f9e9c36c2d89d8a44277033f208a47aa8716e3a19a828d: Status 404 returned error can't find the container with id 79f0b515be9b1050e7f9e9c36c2d89d8a44277033f208a47aa8716e3a19a828d Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.749700 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qlhc\" (UniqueName: \"kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc\") pod \"certified-operators-8jnz2\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.791588 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.791777 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.29175228 +0000 UTC m=+171.582588000 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.791866 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.792077 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.792250 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.824930 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.829290 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.848658 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:08:38 crc kubenswrapper[4730]: W0221 00:08:38.868812 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0f69aa77_802a_47fe_8367_06a3cce5a134.slice/crio-b4a3978fb98e0fde68023574a0e02c34e8b039493443797cd783096f76597f45 WatchSource:0}: Error finding container b4a3978fb98e0fde68023574a0e02c34e8b039493443797cd783096f76597f45: Status 404 returned error can't find the container with id b4a3978fb98e0fde68023574a0e02c34e8b039493443797cd783096f76597f45 Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.890110 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerStarted","Data":"79f0b515be9b1050e7f9e9c36c2d89d8a44277033f208a47aa8716e3a19a828d"} Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.891876 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerStarted","Data":"b4a3978fb98e0fde68023574a0e02c34e8b039493443797cd783096f76597f45"} Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.893443 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerStarted","Data":"3972471d1a40ed3de9c1598bafd20ed04872fc57aa083acb76e6889c2d562eb5"} Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.896275 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:38 crc kubenswrapper[4730]: I0221 00:08:38.902005 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:38 crc kubenswrapper[4730]: E0221 00:08:38.902421 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.402367162 +0000 UTC m=+171.693202892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.003303 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.003443 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.503421445 +0000 UTC m=+171.794257185 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.003836 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.004326 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.504318451 +0000 UTC m=+171.795154181 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.105624 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.105796 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.605770175 +0000 UTC m=+171.896605905 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.106041 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.106349 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.606335162 +0000 UTC m=+171.897170902 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.207622 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.207861 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.707828797 +0000 UTC m=+171.998664537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.208132 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.208718 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.708700723 +0000 UTC m=+171.999536463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.219783 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309097 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d65lt\" (UniqueName: \"kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt\") pod \"ad7385c1-1a71-45ee-9758-d633abe871f2\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309177 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume\") pod \"ad7385c1-1a71-45ee-9758-d633abe871f2\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309210 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309281 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309340 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume\") pod \"ad7385c1-1a71-45ee-9758-d633abe871f2\" (UID: \"ad7385c1-1a71-45ee-9758-d633abe871f2\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.309411 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.809388195 +0000 UTC m=+172.100224025 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.309971 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.310094 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume" (OuterVolumeSpecName: "config-volume") pod "ad7385c1-1a71-45ee-9758-d633abe871f2" (UID: "ad7385c1-1a71-45ee-9758-d633abe871f2"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.310343 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.810331212 +0000 UTC m=+172.101166942 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.324258 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ad7385c1-1a71-45ee-9758-d633abe871f2" (UID: "ad7385c1-1a71-45ee-9758-d633abe871f2"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.331785 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt" (OuterVolumeSpecName: "kube-api-access-d65lt") pod "ad7385c1-1a71-45ee-9758-d633abe871f2" (UID: "ad7385c1-1a71-45ee-9758-d633abe871f2"). InnerVolumeSpecName "kube-api-access-d65lt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.332793 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-444fb" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.337059 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.347108 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:39 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:39 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:39 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.347160 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.422529 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.422857 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.92283253 +0000 UTC m=+172.213668260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.423228 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.423310 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d65lt\" (UniqueName: \"kubernetes.io/projected/ad7385c1-1a71-45ee-9758-d633abe871f2-kube-api-access-d65lt\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.423326 4730 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ad7385c1-1a71-45ee-9758-d633abe871f2-config-volume\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.423339 4730 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ad7385c1-1a71-45ee-9758-d633abe871f2-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.423648 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:39.923628803 +0000 UTC m=+172.214464533 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.428917 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.428956 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.439120 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.439778 4730 patch_prober.go:28] interesting pod/console-f9d7485db-mk5nk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.439841 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-mk5nk" podUID="d34ad48a-e9a9-47dc-a4b2-16edb864a63a" containerName="console" probeResult="failure" output="Get \"https://10.217.0.18:8443/health\": dial tcp 10.217.0.18:8443: connect: connection refused" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.488578 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.498064 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.498132 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.524087 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.524868 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.02483405 +0000 UTC m=+172.315669870 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.627030 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.127014916 +0000 UTC m=+172.417850636 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.628344 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.636528 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-ht8kr" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.691564 4730 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fec200f_1602_4e98_a86a_9f1d8d633ccb.slice/crio-fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fec200f_1602_4e98_a86a_9f1d8d633ccb.slice/crio-conmon-fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd.scope\": RecentStats: unable to find data in memory cache]" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.731037 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.731348 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.231314103 +0000 UTC m=+172.522149823 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.732624 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.733385 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.233217489 +0000 UTC m=+172.524053219 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.834265 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.835212 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.335174408 +0000 UTC m=+172.626010138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.917188 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.917581 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad7385c1-1a71-45ee-9758-d633abe871f2" containerName="collect-profiles" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.917609 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad7385c1-1a71-45ee-9758-d633abe871f2" containerName="collect-profiles" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.917833 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad7385c1-1a71-45ee-9758-d633abe871f2" containerName="collect-profiles" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.921186 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.936070 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:39 crc kubenswrapper[4730]: E0221 00:08:39.936442 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.436426807 +0000 UTC m=+172.727262547 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.944876 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.965613 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.975940 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.978694 4730 generic.go:334] "Generic (PLEG): container finished" podID="17855419-b9b5-4c88-a004-694eb8320ea7" containerID="10110d5987eddb8dfbf41b31795b16aa36eff1fee09ea3651f5190afd20fcba6" exitCode=0 Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.978953 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerDied","Data":"10110d5987eddb8dfbf41b31795b16aa36eff1fee09ea3651f5190afd20fcba6"} Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.983812 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-zwmzp" Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.984340 4730 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.988099 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" event={"ID":"2baa63ee-4f71-41e4-830f-9589b69e1faa","Type":"ContainerStarted","Data":"d92676c5881acc78c567b78f25fbc17e082eebf09b0126771e19cb6a70f487cf"} Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.990316 4730 generic.go:334] "Generic (PLEG): container finished" podID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerID="fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd" exitCode=0 Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.990370 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerDied","Data":"fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd"} Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.990392 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerStarted","Data":"bc7bcc1e3812adcc0ac62d7802930d1d9f6e6908426c81ba11cee27ae65b4e00"} Feb 21 00:08:39 crc kubenswrapper[4730]: I0221 00:08:39.994430 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892","Type":"ContainerStarted","Data":"1cf2e56b1249e55a9ca6f48ca330685b384f9a733bac69d196a44034ecac3e7d"} Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:39.998810 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" event={"ID":"ad7385c1-1a71-45ee-9758-d633abe871f2","Type":"ContainerDied","Data":"0093fdb1f4bb3759a4f6012da44fa0bbce94e2bd5cd1dd4f7b1629ce691b3e55"} Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:39.998851 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0093fdb1f4bb3759a4f6012da44fa0bbce94e2bd5cd1dd4f7b1629ce691b3e55" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:39.998959 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527200-frqmr" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.010766 4730 generic.go:334] "Generic (PLEG): container finished" podID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerID="9d4f5a2a9be586fcd8f0e65cf0835f92dc5b6f3692f16212dda3b235c49e2b57" exitCode=0 Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.010827 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerDied","Data":"9d4f5a2a9be586fcd8f0e65cf0835f92dc5b6f3692f16212dda3b235c49e2b57"} Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.017096 4730 generic.go:334] "Generic (PLEG): container finished" podID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerID="214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168" exitCode=0 Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.017137 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerDied","Data":"214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168"} Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.036926 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.037036 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.537010576 +0000 UTC m=+172.827846296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.037217 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.037308 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.037349 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.037371 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s62rx\" (UniqueName: \"kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.037580 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.537564932 +0000 UTC m=+172.828400662 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.139719 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.140104 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.140154 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.140181 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s62rx\" (UniqueName: \"kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.140764 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.640743677 +0000 UTC m=+172.931579407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.141814 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.145163 4730 patch_prober.go:28] interesting pod/apiserver-76f77b778f-m9mhf container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]log ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]etcd ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/start-apiserver-admission-initializer ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/generic-apiserver-start-informers ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/max-in-flight-filter ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/storage-object-count-tracker-hook ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/image.openshift.io-apiserver-caches ok Feb 21 00:08:40 crc kubenswrapper[4730]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Feb 21 00:08:40 crc kubenswrapper[4730]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/project.openshift.io-projectcache ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/openshift.io-startinformers ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/openshift.io-restmapperupdater ok Feb 21 00:08:40 crc kubenswrapper[4730]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Feb 21 00:08:40 crc kubenswrapper[4730]: livez check failed Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.145240 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" podUID="2d012ccd-e3fb-4dbb-b375-e70f9d803919" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.152413 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.201372 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s62rx\" (UniqueName: \"kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx\") pod \"redhat-marketplace-5hhj6\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.243222 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.243759 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.743743356 +0000 UTC m=+173.034579086 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.271601 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.295138 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.295577 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.295138 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.295791 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.330631 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.331640 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.336436 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.341564 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:40 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:40 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:40 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.341632 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.344397 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.344831 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.844816978 +0000 UTC m=+173.135652708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.406999 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.409981 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.424700 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-qnt7r" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.446290 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.446335 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdlhw\" (UniqueName: \"kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.446420 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.446440 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.446779 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:40.946766457 +0000 UTC m=+173.237602187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.544470 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.547291 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.547518 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-02-21 00:08:41.04748734 +0000 UTC m=+173.338323160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.547605 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdlhw\" (UniqueName: \"kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.547841 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.547874 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.549448 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.549576 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: W0221 00:08:40.557487 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod063a7dbe_6af4_4aff_9d3a_6f4394b98366.slice/crio-222f3f7bd6c6e2f97c2bfa1d995846f8032a2fa1c8664c2bbc2409cf5b245a03 WatchSource:0}: Error finding container 222f3f7bd6c6e2f97c2bfa1d995846f8032a2fa1c8664c2bbc2409cf5b245a03: Status 404 returned error can't find the container with id 222f3f7bd6c6e2f97c2bfa1d995846f8032a2fa1c8664c2bbc2409cf5b245a03 Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.571380 4730 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.573087 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdlhw\" (UniqueName: \"kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw\") pod \"redhat-marketplace-znbtz\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.649555 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: E0221 00:08:40.650007 4730 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-02-21 00:08:41.149986315 +0000 UTC m=+173.440822065 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-nnhll" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.654055 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.680777 4730 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-02-21T00:08:40.571412489Z","Handler":null,"Name":""} Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.687423 4730 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.687576 4730 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.746308 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.747277 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.751304 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.751473 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.754346 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.758450 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.761784 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.852922 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.853278 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.853767 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.865138 4730 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.865171 4730 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.902235 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-nnhll\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.954881 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.955062 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.955245 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:40 crc kubenswrapper[4730]: I0221 00:08:40.976368 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.031827 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" event={"ID":"2baa63ee-4f71-41e4-830f-9589b69e1faa","Type":"ContainerStarted","Data":"6a0f54eb38425381d43e8c7bcd9776c787e8c5f37c16a0b1305396abd27f4c6f"} Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.031892 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" event={"ID":"2baa63ee-4f71-41e4-830f-9589b69e1faa","Type":"ContainerStarted","Data":"a77a4ea9b1a49c350895d92b09358d5f057c745a302d2e7ac2a2e6d397ca58f3"} Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.035674 4730 generic.go:334] "Generic (PLEG): container finished" podID="7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" containerID="b603493b55367653b3a54b8fe696ba503009d5d45822b06e5ed61daece32168b" exitCode=0 Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.035853 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892","Type":"ContainerDied","Data":"b603493b55367653b3a54b8fe696ba503009d5d45822b06e5ed61daece32168b"} Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.041085 4730 generic.go:334] "Generic (PLEG): container finished" podID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerID="7c2b7d1f20613c9af58bd7dae231689b8e72c73cdeea040e7f0ed220babf5798" exitCode=0 Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.041220 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerDied","Data":"7c2b7d1f20613c9af58bd7dae231689b8e72c73cdeea040e7f0ed220babf5798"} Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.041310 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerStarted","Data":"222f3f7bd6c6e2f97c2bfa1d995846f8032a2fa1c8664c2bbc2409cf5b245a03"} Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.061787 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-n66hq" podStartSLOduration=14.061767167 podStartE2EDuration="14.061767167s" podCreationTimestamp="2026-02-21 00:08:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:41.057664538 +0000 UTC m=+173.348500288" watchObservedRunningTime="2026-02-21 00:08:41.061767167 +0000 UTC m=+173.352602897" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.072777 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.107416 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.145238 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.309404 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.311835 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.317642 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.318033 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.341157 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:41 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:41 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:41 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.341202 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.373728 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Feb 21 00:08:41 crc kubenswrapper[4730]: W0221 00:08:41.395271 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode68aa6d1_579b_4abe_8670_14231fb2853e.slice/crio-79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1 WatchSource:0}: Error finding container 79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1: Status 404 returned error can't find the container with id 79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1 Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.439682 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:08:41 crc kubenswrapper[4730]: W0221 00:08:41.454657 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod691c1af1_e7b5_4b53_8f4d_9a33e48106be.slice/crio-97a7852ad22c32f44bbc120d697bfcdbaa504a08dec964c49cc1a7b55f2aed02 WatchSource:0}: Error finding container 97a7852ad22c32f44bbc120d697bfcdbaa504a08dec964c49cc1a7b55f2aed02: Status 404 returned error can't find the container with id 97a7852ad22c32f44bbc120d697bfcdbaa504a08dec964c49cc1a7b55f2aed02 Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.464749 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.464954 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj4zb\" (UniqueName: \"kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.465054 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.566363 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.566823 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj4zb\" (UniqueName: \"kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.566863 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.567087 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.567229 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.586557 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj4zb\" (UniqueName: \"kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb\") pod \"redhat-operators-4j2h7\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.636734 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.718920 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.719868 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.727389 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.856861 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.872172 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.872259 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.872295 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wlng\" (UniqueName: \"kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.975995 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.976117 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.976156 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wlng\" (UniqueName: \"kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.977164 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:41 crc kubenswrapper[4730]: I0221 00:08:41.979840 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.018824 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wlng\" (UniqueName: \"kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng\") pod \"redhat-operators-4bmlw\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.046041 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.049490 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerStarted","Data":"fdf65f079d4e0207471013613fdb0235d927ef1c9c284a5ed82ed544712e9f8b"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.052417 4730 generic.go:334] "Generic (PLEG): container finished" podID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerID="aa068a1421584a957dfd93624e21bf4c9adcccfa068353cfbc124a3ada16d734" exitCode=0 Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.052492 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerDied","Data":"aa068a1421584a957dfd93624e21bf4c9adcccfa068353cfbc124a3ada16d734"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.052512 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerStarted","Data":"5e8590b10d4fae4e4284ac206eda32330bdaab27f522820584e6d9c92bc37d99"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.060989 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e68aa6d1-579b-4abe-8670-14231fb2853e","Type":"ContainerStarted","Data":"df022d46730c1a2d117016bdfcd20d4f00bf52ef5a9963e660c0e4964eedd47f"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.061036 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e68aa6d1-579b-4abe-8670-14231fb2853e","Type":"ContainerStarted","Data":"79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.064610 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" event={"ID":"691c1af1-e7b5-4b53-8f4d-9a33e48106be","Type":"ContainerStarted","Data":"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.064646 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.064657 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" event={"ID":"691c1af1-e7b5-4b53-8f4d-9a33e48106be","Type":"ContainerStarted","Data":"97a7852ad22c32f44bbc120d697bfcdbaa504a08dec964c49cc1a7b55f2aed02"} Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.089804 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" podStartSLOduration=155.089786155 podStartE2EDuration="2m35.089786155s" podCreationTimestamp="2026-02-21 00:06:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:42.089056464 +0000 UTC m=+174.379892194" watchObservedRunningTime="2026-02-21 00:08:42.089786155 +0000 UTC m=+174.380621885" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.101548 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=2.101528929 podStartE2EDuration="2.101528929s" podCreationTimestamp="2026-02-21 00:08:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:08:42.101126247 +0000 UTC m=+174.391961977" watchObservedRunningTime="2026-02-21 00:08:42.101528929 +0000 UTC m=+174.392364659" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.341673 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:42 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:42 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:42 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.342058 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.442280 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.442793 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.486033 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.597804 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir\") pod \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.598321 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access\") pod \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\" (UID: \"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892\") " Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.598055 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" (UID: "7ed8a507-e3ec-4e06-b0c0-2ff5283b8892"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.603769 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" (UID: "7ed8a507-e3ec-4e06-b0c0-2ff5283b8892"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.700494 4730 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:42 crc kubenswrapper[4730]: I0221 00:08:42.700522 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7ed8a507-e3ec-4e06-b0c0-2ff5283b8892-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.089526 4730 generic.go:334] "Generic (PLEG): container finished" podID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerID="20ec430915a54169ae28cee1bd4eb61951bc732541b2f3d339c4c6b8ec7e3861" exitCode=0 Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.089630 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerDied","Data":"20ec430915a54169ae28cee1bd4eb61951bc732541b2f3d339c4c6b8ec7e3861"} Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.092691 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.093129 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"7ed8a507-e3ec-4e06-b0c0-2ff5283b8892","Type":"ContainerDied","Data":"1cf2e56b1249e55a9ca6f48ca330685b384f9a733bac69d196a44034ecac3e7d"} Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.093191 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1cf2e56b1249e55a9ca6f48ca330685b384f9a733bac69d196a44034ecac3e7d" Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.095202 4730 generic.go:334] "Generic (PLEG): container finished" podID="e68aa6d1-579b-4abe-8670-14231fb2853e" containerID="df022d46730c1a2d117016bdfcd20d4f00bf52ef5a9963e660c0e4964eedd47f" exitCode=0 Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.095265 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e68aa6d1-579b-4abe-8670-14231fb2853e","Type":"ContainerDied","Data":"df022d46730c1a2d117016bdfcd20d4f00bf52ef5a9963e660c0e4964eedd47f"} Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.099503 4730 generic.go:334] "Generic (PLEG): container finished" podID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerID="42dca4c184df218d49bcc40db386e42ed66a3d0f2d83223200835e289cec8206" exitCode=0 Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.099602 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerDied","Data":"42dca4c184df218d49bcc40db386e42ed66a3d0f2d83223200835e289cec8206"} Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.099681 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerStarted","Data":"24d75dad387611f0daedb58d6ca34943414bb4509e8bd50a5af52baaffd6718c"} Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.346378 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:43 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:43 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:43 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:43 crc kubenswrapper[4730]: I0221 00:08:43.346667 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.340076 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:44 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:44 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:44 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.340165 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.411227 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.505510 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.510893 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-m9mhf" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.572772 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir\") pod \"e68aa6d1-579b-4abe-8670-14231fb2853e\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.572885 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access\") pod \"e68aa6d1-579b-4abe-8670-14231fb2853e\" (UID: \"e68aa6d1-579b-4abe-8670-14231fb2853e\") " Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.572926 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e68aa6d1-579b-4abe-8670-14231fb2853e" (UID: "e68aa6d1-579b-4abe-8670-14231fb2853e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.573158 4730 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e68aa6d1-579b-4abe-8670-14231fb2853e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.580735 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e68aa6d1-579b-4abe-8670-14231fb2853e" (UID: "e68aa6d1-579b-4abe-8670-14231fb2853e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:08:44 crc kubenswrapper[4730]: I0221 00:08:44.674698 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e68aa6d1-579b-4abe-8670-14231fb2853e-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.128323 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e68aa6d1-579b-4abe-8670-14231fb2853e","Type":"ContainerDied","Data":"79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1"} Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.128372 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.128393 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79f77c584fe73b32cfdf488d92d73aab9638369dc86be2f22ba8a56925fa16d1" Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.340635 4730 patch_prober.go:28] interesting pod/router-default-5444994796-4wrmp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Feb 21 00:08:45 crc kubenswrapper[4730]: [-]has-synced failed: reason withheld Feb 21 00:08:45 crc kubenswrapper[4730]: [+]process-running ok Feb 21 00:08:45 crc kubenswrapper[4730]: healthz check failed Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.341029 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-4wrmp" podUID="0d89fc71-51e8-4595-b9d4-fc0cd8103017" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Feb 21 00:08:45 crc kubenswrapper[4730]: I0221 00:08:45.509818 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-djs5z" Feb 21 00:08:46 crc kubenswrapper[4730]: I0221 00:08:46.343547 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:46 crc kubenswrapper[4730]: I0221 00:08:46.353358 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-4wrmp" Feb 21 00:08:49 crc kubenswrapper[4730]: I0221 00:08:49.433188 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:49 crc kubenswrapper[4730]: I0221 00:08:49.436404 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-mk5nk" Feb 21 00:08:50 crc kubenswrapper[4730]: I0221 00:08:50.297929 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:50 crc kubenswrapper[4730]: I0221 00:08:50.298478 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:08:50 crc kubenswrapper[4730]: I0221 00:08:50.298513 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:50 crc kubenswrapper[4730]: I0221 00:08:50.298495 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:08:55 crc kubenswrapper[4730]: I0221 00:08:55.613641 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:08:55 crc kubenswrapper[4730]: I0221 00:08:55.614542 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" containerID="cri-o://15bdb7ede88d5b9da751da95e920fa2ddb139ffba22a0255d6705009aacd8f09" gracePeriod=30 Feb 21 00:08:55 crc kubenswrapper[4730]: I0221 00:08:55.628644 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:08:55 crc kubenswrapper[4730]: I0221 00:08:55.628966 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" containerID="cri-o://3f694844b8366bd91f34a6a1fb0cc86b4182dd1145cac0b321e62387d5297478" gracePeriod=30 Feb 21 00:08:56 crc kubenswrapper[4730]: I0221 00:08:56.231792 4730 generic.go:334] "Generic (PLEG): container finished" podID="0804eedb-db79-4012-85c6-f3e572f93179" containerID="15bdb7ede88d5b9da751da95e920fa2ddb139ffba22a0255d6705009aacd8f09" exitCode=0 Feb 21 00:08:56 crc kubenswrapper[4730]: I0221 00:08:56.231882 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" event={"ID":"0804eedb-db79-4012-85c6-f3e572f93179","Type":"ContainerDied","Data":"15bdb7ede88d5b9da751da95e920fa2ddb139ffba22a0255d6705009aacd8f09"} Feb 21 00:08:56 crc kubenswrapper[4730]: I0221 00:08:56.234286 4730 generic.go:334] "Generic (PLEG): container finished" podID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerID="3f694844b8366bd91f34a6a1fb0cc86b4182dd1145cac0b321e62387d5297478" exitCode=0 Feb 21 00:08:56 crc kubenswrapper[4730]: I0221 00:08:56.234363 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" event={"ID":"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf","Type":"ContainerDied","Data":"3f694844b8366bd91f34a6a1fb0cc86b4182dd1145cac0b321e62387d5297478"} Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.296639 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.296663 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.296983 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297015 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297039 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297494 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297523 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297685 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"691135f2d846b50187897515760760f57da831222faa7ad2a02c133c5ab2c08d"} pod="openshift-console/downloads-7954f5f757-bk7rn" containerMessage="Container download-server failed liveness probe, will be restarted" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.297766 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" containerID="cri-o://691135f2d846b50187897515760760f57da831222faa7ad2a02c133c5ab2c08d" gracePeriod=2 Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.486268 4730 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qjdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.486332 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.557461 4730 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lj2pw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 21 00:09:00 crc kubenswrapper[4730]: I0221 00:09:00.557573 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 21 00:09:01 crc kubenswrapper[4730]: I0221 00:09:01.114103 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:09:01 crc kubenswrapper[4730]: I0221 00:09:01.281599 4730 generic.go:334] "Generic (PLEG): container finished" podID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerID="691135f2d846b50187897515760760f57da831222faa7ad2a02c133c5ab2c08d" exitCode=0 Feb 21 00:09:01 crc kubenswrapper[4730]: I0221 00:09:01.281649 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bk7rn" event={"ID":"0fbf3e59-6719-4583-a0aa-4badc34ecf3b","Type":"ContainerDied","Data":"691135f2d846b50187897515760760f57da831222faa7ad2a02c133c5ab2c08d"} Feb 21 00:09:03 crc kubenswrapper[4730]: E0221 00:09:03.198369 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 21 00:09:03 crc kubenswrapper[4730]: E0221 00:09:03.198511 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-l9qmc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-pjxx6_openshift-marketplace(0f69aa77-802a-47fe-8367-06a3cce5a134): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 21 00:09:03 crc kubenswrapper[4730]: E0221 00:09:03.199723 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-pjxx6" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" Feb 21 00:09:05 crc kubenswrapper[4730]: I0221 00:09:05.775405 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:09:05 crc kubenswrapper[4730]: I0221 00:09:05.775665 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:09:07 crc kubenswrapper[4730]: I0221 00:09:07.313486 4730 generic.go:334] "Generic (PLEG): container finished" podID="05c8cb3c-a432-4eef-895d-6154a07f3b90" containerID="1808902e42f592dbdd29c600d68892edf306e266e013adc83dd1355b27867da2" exitCode=0 Feb 21 00:09:07 crc kubenswrapper[4730]: I0221 00:09:07.313587 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29527200-hgrnz" event={"ID":"05c8cb3c-a432-4eef-895d-6154a07f3b90","Type":"ContainerDied","Data":"1808902e42f592dbdd29c600d68892edf306e266e013adc83dd1355b27867da2"} Feb 21 00:09:09 crc kubenswrapper[4730]: E0221 00:09:09.441176 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-pjxx6" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.295788 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.295883 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.389201 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m5k9h" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.485145 4730 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-2qjdj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.485258 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.559211 4730 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-lj2pw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.559291 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Feb 21 00:09:10 crc kubenswrapper[4730]: E0221 00:09:10.582391 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Feb 21 00:09:10 crc kubenswrapper[4730]: E0221 00:09:10.583143 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gdlhw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-znbtz_openshift-marketplace(e68cc82a-3bb4-40df-bab5-a9b4fa735bbb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 21 00:09:10 crc kubenswrapper[4730]: E0221 00:09:10.584552 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-znbtz" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.637862 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.696145 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca\") pod \"05c8cb3c-a432-4eef-895d-6154a07f3b90\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.696372 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8285\" (UniqueName: \"kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285\") pod \"05c8cb3c-a432-4eef-895d-6154a07f3b90\" (UID: \"05c8cb3c-a432-4eef-895d-6154a07f3b90\") " Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.697781 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca" (OuterVolumeSpecName: "serviceca") pod "05c8cb3c-a432-4eef-895d-6154a07f3b90" (UID: "05c8cb3c-a432-4eef-895d-6154a07f3b90"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.706413 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285" (OuterVolumeSpecName: "kube-api-access-k8285") pod "05c8cb3c-a432-4eef-895d-6154a07f3b90" (UID: "05c8cb3c-a432-4eef-895d-6154a07f3b90"). InnerVolumeSpecName "kube-api-access-k8285". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.798461 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8285\" (UniqueName: \"kubernetes.io/projected/05c8cb3c-a432-4eef-895d-6154a07f3b90-kube-api-access-k8285\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:10 crc kubenswrapper[4730]: I0221 00:09:10.798515 4730 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/05c8cb3c-a432-4eef-895d-6154a07f3b90-serviceca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:11 crc kubenswrapper[4730]: I0221 00:09:11.340528 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29527200-hgrnz" Feb 21 00:09:11 crc kubenswrapper[4730]: I0221 00:09:11.340664 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29527200-hgrnz" event={"ID":"05c8cb3c-a432-4eef-895d-6154a07f3b90","Type":"ContainerDied","Data":"afc839afa37bcbd02461a8f1e448c6cb7df96cc8b8cb0fb4ad7f6c1d155a05ae"} Feb 21 00:09:11 crc kubenswrapper[4730]: I0221 00:09:11.340744 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="afc839afa37bcbd02461a8f1e448c6cb7df96cc8b8cb0fb4ad7f6c1d155a05ae" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.287225 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-znbtz" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.351940 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" event={"ID":"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf","Type":"ContainerDied","Data":"40aead9276bffd87d97236d10c5dcc3450f0f40eb98c4e11687509f65707b7cb"} Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.351985 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40aead9276bffd87d97236d10c5dcc3450f0f40eb98c4e11687509f65707b7cb" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.354369 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" event={"ID":"0804eedb-db79-4012-85c6-f3e572f93179","Type":"ContainerDied","Data":"f7f916ac8b6ec59ecd004b9d06b9f3bec6d4ec078cc8d5c63b17dff3848b8e2f"} Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.354420 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7f916ac8b6ec59ecd004b9d06b9f3bec6d4ec078cc8d5c63b17dff3848b8e2f" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.371048 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.379122 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.430010 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.430284 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8vtzm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rk2d5_openshift-marketplace(d8b1b254-42f2-4ff9-8971-13528ab4f8d0): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.430876 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.431086 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6qlhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-8jnz2_openshift-marketplace(5fec200f-1602-4e98-a86a-9f1d8d633ccb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.431515 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rk2d5" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.435310 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-8jnz2" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.437542 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.437680 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-drr9w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-d4xfg_openshift-marketplace(17855419-b9b5-4c88-a004-694eb8320ea7): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.438819 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-d4xfg" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.531590 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert\") pod \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.531891 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config\") pod \"0804eedb-db79-4012-85c6-f3e572f93179\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.531948 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config\") pod \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.531972 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles\") pod \"0804eedb-db79-4012-85c6-f3e572f93179\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532007 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca\") pod \"0804eedb-db79-4012-85c6-f3e572f93179\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532093 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jgssk\" (UniqueName: \"kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk\") pod \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532117 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert\") pod \"0804eedb-db79-4012-85c6-f3e572f93179\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532158 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpcjv\" (UniqueName: \"kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv\") pod \"0804eedb-db79-4012-85c6-f3e572f93179\" (UID: \"0804eedb-db79-4012-85c6-f3e572f93179\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532179 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca\") pod \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\" (UID: \"9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf\") " Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.532740 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config" (OuterVolumeSpecName: "config") pod "0804eedb-db79-4012-85c6-f3e572f93179" (UID: "0804eedb-db79-4012-85c6-f3e572f93179"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.533030 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca" (OuterVolumeSpecName: "client-ca") pod "0804eedb-db79-4012-85c6-f3e572f93179" (UID: "0804eedb-db79-4012-85c6-f3e572f93179"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.533095 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config" (OuterVolumeSpecName: "config") pod "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" (UID: "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.533103 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca" (OuterVolumeSpecName: "client-ca") pod "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" (UID: "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.533288 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "0804eedb-db79-4012-85c6-f3e572f93179" (UID: "0804eedb-db79-4012-85c6-f3e572f93179"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.538343 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" (UID: "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.538503 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk" (OuterVolumeSpecName: "kube-api-access-jgssk") pod "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" (UID: "9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf"). InnerVolumeSpecName "kube-api-access-jgssk". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.538607 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv" (OuterVolumeSpecName: "kube-api-access-wpcjv") pod "0804eedb-db79-4012-85c6-f3e572f93179" (UID: "0804eedb-db79-4012-85c6-f3e572f93179"). InnerVolumeSpecName "kube-api-access-wpcjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.539461 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0804eedb-db79-4012-85c6-f3e572f93179" (UID: "0804eedb-db79-4012-85c6-f3e572f93179"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633772 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633822 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jgssk\" (UniqueName: \"kubernetes.io/projected/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-kube-api-access-jgssk\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633837 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0804eedb-db79-4012-85c6-f3e572f93179-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633852 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpcjv\" (UniqueName: \"kubernetes.io/projected/0804eedb-db79-4012-85c6-f3e572f93179-kube-api-access-wpcjv\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633863 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633875 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633886 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633913 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.633993 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0804eedb-db79-4012-85c6-f3e572f93179-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977038 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.977428 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977454 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.977474 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977486 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.977503 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68aa6d1-579b-4abe-8670-14231fb2853e" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977516 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68aa6d1-579b-4abe-8670-14231fb2853e" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.977541 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05c8cb3c-a432-4eef-895d-6154a07f3b90" containerName="image-pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977553 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="05c8cb3c-a432-4eef-895d-6154a07f3b90" containerName="image-pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: E0221 00:09:12.977581 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977594 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977794 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="05c8cb3c-a432-4eef-895d-6154a07f3b90" containerName="image-pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977818 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68aa6d1-579b-4abe-8670-14231fb2853e" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977858 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" containerName="route-controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977880 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="0804eedb-db79-4012-85c6-f3e572f93179" containerName="controller-manager" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.977952 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ed8a507-e3ec-4e06-b0c0-2ff5283b8892" containerName="pruner" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.978556 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.985477 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.986427 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:12 crc kubenswrapper[4730]: I0221 00:09:12.993604 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.000054 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.141693 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r7wht\" (UniqueName: \"kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142038 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142064 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142094 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142113 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142128 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142162 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9g4w6\" (UniqueName: \"kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142453 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.142572 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243349 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243413 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243430 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243463 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9g4w6\" (UniqueName: \"kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243513 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243552 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243578 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r7wht\" (UniqueName: \"kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243598 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.243614 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.244830 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.244938 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.244960 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.245426 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.247805 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.248423 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.250940 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.263520 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9g4w6\" (UniqueName: \"kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6\") pod \"controller-manager-f7d9f84f5-d9ccg\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.263917 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r7wht\" (UniqueName: \"kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht\") pod \"route-controller-manager-7fb457f777-9x9q6\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.310102 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.324448 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.363153 4730 generic.go:334] "Generic (PLEG): container finished" podID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerID="91de3cddfcbe8b1005c4de6537c6ca655f7008d41555ab3a3813e4f2268cee28" exitCode=0 Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.363264 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerDied","Data":"91de3cddfcbe8b1005c4de6537c6ca655f7008d41555ab3a3813e4f2268cee28"} Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.368435 4730 generic.go:334] "Generic (PLEG): container finished" podID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerID="d7ed0e0b8e21b25cece0a7f7e047cc5aa423619f2090c38ff80cf067c283736c" exitCode=0 Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.368504 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerDied","Data":"d7ed0e0b8e21b25cece0a7f7e047cc5aa423619f2090c38ff80cf067c283736c"} Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.372229 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-bk7rn" event={"ID":"0fbf3e59-6719-4583-a0aa-4badc34ecf3b","Type":"ContainerStarted","Data":"87b617f96c5d507d249aed4a25245f48f9f70da3fb6a5061af4e3731b4b79fb8"} Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.373179 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.373262 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.373292 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.375057 4730 generic.go:334] "Generic (PLEG): container finished" podID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerID="623f80859bfcb4356487021bf1a827e996e7ddd6a14c82ab46345c76f31c0805" exitCode=0 Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.375863 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerDied","Data":"623f80859bfcb4356487021bf1a827e996e7ddd6a14c82ab46345c76f31c0805"} Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.376272 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw" Feb 21 00:09:13 crc kubenswrapper[4730]: E0221 00:09:13.377808 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rk2d5" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" Feb 21 00:09:13 crc kubenswrapper[4730]: E0221 00:09:13.378040 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-d4xfg" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.378168 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-2qjdj" Feb 21 00:09:13 crc kubenswrapper[4730]: E0221 00:09:13.386343 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-8jnz2" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.571394 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.579131 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-2qjdj"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.591990 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.594086 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-lj2pw"] Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.607997 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:13 crc kubenswrapper[4730]: W0221 00:09:13.616044 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47d2efaf_b6ec_4250_803d_0ddfe2963bc3.slice/crio-385c10fc8ae7a80de17caafc986ca2c0093124e3b681a63d6873425f83226e56 WatchSource:0}: Error finding container 385c10fc8ae7a80de17caafc986ca2c0093124e3b681a63d6873425f83226e56: Status 404 returned error can't find the container with id 385c10fc8ae7a80de17caafc986ca2c0093124e3b681a63d6873425f83226e56 Feb 21 00:09:13 crc kubenswrapper[4730]: I0221 00:09:13.857721 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:13 crc kubenswrapper[4730]: W0221 00:09:13.874461 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podddb567d7_d6a7_4c5d_a54a_515098f2cce5.slice/crio-36883b745dfc6e9105f6814a984c625dd4ac106bbb2cec3ce7af873472b56d2c WatchSource:0}: Error finding container 36883b745dfc6e9105f6814a984c625dd4ac106bbb2cec3ce7af873472b56d2c: Status 404 returned error can't find the container with id 36883b745dfc6e9105f6814a984c625dd4ac106bbb2cec3ce7af873472b56d2c Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.381293 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" event={"ID":"47d2efaf-b6ec-4250-803d-0ddfe2963bc3","Type":"ContainerStarted","Data":"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.381746 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.381759 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" event={"ID":"47d2efaf-b6ec-4250-803d-0ddfe2963bc3","Type":"ContainerStarted","Data":"385c10fc8ae7a80de17caafc986ca2c0093124e3b681a63d6873425f83226e56"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.383884 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerStarted","Data":"a9d7dd6348892cd6c3793c2ee126f7b668e62b21b734935b136b35c8549a9cab"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.385883 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerStarted","Data":"f0b3831ab0b1167b7f2b6721b728e79e167c63cc1825444be22b8327f03f16a1"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.387869 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" event={"ID":"ddb567d7-d6a7-4c5d-a54a-515098f2cce5","Type":"ContainerStarted","Data":"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.387895 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" event={"ID":"ddb567d7-d6a7-4c5d-a54a-515098f2cce5","Type":"ContainerStarted","Data":"36883b745dfc6e9105f6814a984c625dd4ac106bbb2cec3ce7af873472b56d2c"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.388398 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.391140 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.391172 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.391394 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerStarted","Data":"b1924224a57f8dd76b64e50698b0a6fdbb722fc96e57b5460ff8e68e3c7dd8a1"} Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.395029 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.400444 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.403360 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" podStartSLOduration=19.403351849 podStartE2EDuration="19.403351849s" podCreationTimestamp="2026-02-21 00:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:14.400834696 +0000 UTC m=+206.691670426" watchObservedRunningTime="2026-02-21 00:09:14.403351849 +0000 UTC m=+206.694187579" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.421032 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bmlw" podStartSLOduration=2.717801223 podStartE2EDuration="33.421010872s" podCreationTimestamp="2026-02-21 00:08:41 +0000 UTC" firstStartedPulling="2026-02-21 00:08:43.104769893 +0000 UTC m=+175.395605623" lastFinishedPulling="2026-02-21 00:09:13.807979542 +0000 UTC m=+206.098815272" observedRunningTime="2026-02-21 00:09:14.418931652 +0000 UTC m=+206.709767392" watchObservedRunningTime="2026-02-21 00:09:14.421010872 +0000 UTC m=+206.711846602" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.440942 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0804eedb-db79-4012-85c6-f3e572f93179" path="/var/lib/kubelet/pods/0804eedb-db79-4012-85c6-f3e572f93179/volumes" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.442081 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf" path="/var/lib/kubelet/pods/9e501ff1-0e3c-4ba4-84c5-5aa493ef5aaf/volumes" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.442149 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4j2h7" podStartSLOduration=2.6501895170000003 podStartE2EDuration="33.442129825s" podCreationTimestamp="2026-02-21 00:08:41 +0000 UTC" firstStartedPulling="2026-02-21 00:08:43.091532736 +0000 UTC m=+175.382368466" lastFinishedPulling="2026-02-21 00:09:13.883473044 +0000 UTC m=+206.174308774" observedRunningTime="2026-02-21 00:09:14.44124792 +0000 UTC m=+206.732083650" watchObservedRunningTime="2026-02-21 00:09:14.442129825 +0000 UTC m=+206.732965555" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.501656 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5hhj6" podStartSLOduration=2.6479521950000002 podStartE2EDuration="35.501625933s" podCreationTimestamp="2026-02-21 00:08:39 +0000 UTC" firstStartedPulling="2026-02-21 00:08:41.04405463 +0000 UTC m=+173.334890370" lastFinishedPulling="2026-02-21 00:09:13.897728378 +0000 UTC m=+206.188564108" observedRunningTime="2026-02-21 00:09:14.496271407 +0000 UTC m=+206.787107127" watchObservedRunningTime="2026-02-21 00:09:14.501625933 +0000 UTC m=+206.792461663" Feb 21 00:09:14 crc kubenswrapper[4730]: I0221 00:09:14.546970 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" podStartSLOduration=19.546951999 podStartE2EDuration="19.546951999s" podCreationTimestamp="2026-02-21 00:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:14.537224626 +0000 UTC m=+206.828060356" watchObservedRunningTime="2026-02-21 00:09:14.546951999 +0000 UTC m=+206.837787729" Feb 21 00:09:15 crc kubenswrapper[4730]: I0221 00:09:15.229043 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:15 crc kubenswrapper[4730]: I0221 00:09:15.342049 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:15 crc kubenswrapper[4730]: I0221 00:09:15.394754 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:15 crc kubenswrapper[4730]: I0221 00:09:15.394818 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:16 crc kubenswrapper[4730]: I0221 00:09:16.406040 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" podUID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" containerName="route-controller-manager" containerID="cri-o://bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115" gracePeriod=30 Feb 21 00:09:16 crc kubenswrapper[4730]: I0221 00:09:16.406337 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" podUID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" containerName="controller-manager" containerID="cri-o://939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac" gracePeriod=30 Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.362498 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.368977 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.390812 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:17 crc kubenswrapper[4730]: E0221 00:09:17.391033 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" containerName="route-controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.391045 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" containerName="route-controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: E0221 00:09:17.391052 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" containerName="controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.391060 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" containerName="controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.391158 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" containerName="route-controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.391172 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" containerName="controller-manager" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.391535 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.405685 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.412677 4730 generic.go:334] "Generic (PLEG): container finished" podID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" containerID="bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115" exitCode=0 Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.412751 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" event={"ID":"47d2efaf-b6ec-4250-803d-0ddfe2963bc3","Type":"ContainerDied","Data":"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115"} Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.412785 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" event={"ID":"47d2efaf-b6ec-4250-803d-0ddfe2963bc3","Type":"ContainerDied","Data":"385c10fc8ae7a80de17caafc986ca2c0093124e3b681a63d6873425f83226e56"} Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.412808 4730 scope.go:117] "RemoveContainer" containerID="bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.412951 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.415235 4730 generic.go:334] "Generic (PLEG): container finished" podID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" containerID="939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac" exitCode=0 Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.415274 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" event={"ID":"ddb567d7-d6a7-4c5d-a54a-515098f2cce5","Type":"ContainerDied","Data":"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac"} Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.415292 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.415304 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg" event={"ID":"ddb567d7-d6a7-4c5d-a54a-515098f2cce5","Type":"ContainerDied","Data":"36883b745dfc6e9105f6814a984c625dd4ac106bbb2cec3ce7af873472b56d2c"} Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.435715 4730 scope.go:117] "RemoveContainer" containerID="bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115" Feb 21 00:09:17 crc kubenswrapper[4730]: E0221 00:09:17.436272 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115\": container with ID starting with bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115 not found: ID does not exist" containerID="bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.436320 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115"} err="failed to get container status \"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115\": rpc error: code = NotFound desc = could not find container \"bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115\": container with ID starting with bd839828c9f3a499e915184cb3944144a34aa463d80fa6184d1bf122d8748115 not found: ID does not exist" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.436343 4730 scope.go:117] "RemoveContainer" containerID="939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.454802 4730 scope.go:117] "RemoveContainer" containerID="939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac" Feb 21 00:09:17 crc kubenswrapper[4730]: E0221 00:09:17.455289 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac\": container with ID starting with 939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac not found: ID does not exist" containerID="939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.455330 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac"} err="failed to get container status \"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac\": rpc error: code = NotFound desc = could not find container \"939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac\": container with ID starting with 939a993550566a1ec19d8bc2b775510f68636258493700d4b33dfc32e797e3ac not found: ID does not exist" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.510779 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca\") pod \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.510853 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r7wht\" (UniqueName: \"kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht\") pod \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.510881 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config\") pod \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.510969 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9g4w6\" (UniqueName: \"kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6\") pod \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.511022 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles\") pod \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.511149 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca\") pod \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.511611 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca" (OuterVolumeSpecName: "client-ca") pod "47d2efaf-b6ec-4250-803d-0ddfe2963bc3" (UID: "47d2efaf-b6ec-4250-803d-0ddfe2963bc3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.511631 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ddb567d7-d6a7-4c5d-a54a-515098f2cce5" (UID: "ddb567d7-d6a7-4c5d-a54a-515098f2cce5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.511654 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca" (OuterVolumeSpecName: "client-ca") pod "ddb567d7-d6a7-4c5d-a54a-515098f2cce5" (UID: "ddb567d7-d6a7-4c5d-a54a-515098f2cce5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512062 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config" (OuterVolumeSpecName: "config") pod "ddb567d7-d6a7-4c5d-a54a-515098f2cce5" (UID: "ddb567d7-d6a7-4c5d-a54a-515098f2cce5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512134 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config\") pod \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512181 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert\") pod \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\" (UID: \"ddb567d7-d6a7-4c5d-a54a-515098f2cce5\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512251 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert\") pod \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\" (UID: \"47d2efaf-b6ec-4250-803d-0ddfe2963bc3\") " Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512459 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512516 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw6x2\" (UniqueName: \"kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512537 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config" (OuterVolumeSpecName: "config") pod "47d2efaf-b6ec-4250-803d-0ddfe2963bc3" (UID: "47d2efaf-b6ec-4250-803d-0ddfe2963bc3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.512767 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513309 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513410 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513437 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513451 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513464 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.513476 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.518633 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht" (OuterVolumeSpecName: "kube-api-access-r7wht") pod "47d2efaf-b6ec-4250-803d-0ddfe2963bc3" (UID: "47d2efaf-b6ec-4250-803d-0ddfe2963bc3"). InnerVolumeSpecName "kube-api-access-r7wht". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.518738 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6" (OuterVolumeSpecName: "kube-api-access-9g4w6") pod "ddb567d7-d6a7-4c5d-a54a-515098f2cce5" (UID: "ddb567d7-d6a7-4c5d-a54a-515098f2cce5"). InnerVolumeSpecName "kube-api-access-9g4w6". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.523990 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "47d2efaf-b6ec-4250-803d-0ddfe2963bc3" (UID: "47d2efaf-b6ec-4250-803d-0ddfe2963bc3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.524027 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ddb567d7-d6a7-4c5d-a54a-515098f2cce5" (UID: "ddb567d7-d6a7-4c5d-a54a-515098f2cce5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.544519 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.545539 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.548092 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.550491 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.552940 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614269 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614345 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw6x2\" (UniqueName: \"kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614401 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614477 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614504 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614537 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614585 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9g4w6\" (UniqueName: \"kubernetes.io/projected/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-kube-api-access-9g4w6\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614600 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ddb567d7-d6a7-4c5d-a54a-515098f2cce5-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614612 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.614623 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r7wht\" (UniqueName: \"kubernetes.io/projected/47d2efaf-b6ec-4250-803d-0ddfe2963bc3-kube-api-access-r7wht\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.615434 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.616029 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.620441 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.632169 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw6x2\" (UniqueName: \"kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2\") pod \"route-controller-manager-5f4db4d774-hmnst\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.712099 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.715118 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.715202 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.715325 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.740376 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.805941 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.816012 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-f7d9f84f5-d9ccg"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.835249 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.840583 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7fb457f777-9x9q6"] Feb 21 00:09:17 crc kubenswrapper[4730]: I0221 00:09:17.873630 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.104996 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.243294 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.426116 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b030658d-1eee-4826-b953-10f7ef16e401","Type":"ContainerStarted","Data":"aa2b94bb5122a7e71df5d01c71be3262214c880ab0338a9e59ee56c1399786b7"} Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.430370 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" event={"ID":"4ff57c21-f26d-4456-b388-e42583659a39","Type":"ContainerStarted","Data":"906fad8c096e4291d3f258d89b566a0f026009745bf079fab8bff1632d88b5b2"} Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.443772 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47d2efaf-b6ec-4250-803d-0ddfe2963bc3" path="/var/lib/kubelet/pods/47d2efaf-b6ec-4250-803d-0ddfe2963bc3/volumes" Feb 21 00:09:18 crc kubenswrapper[4730]: I0221 00:09:18.444599 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ddb567d7-d6a7-4c5d-a54a-515098f2cce5" path="/var/lib/kubelet/pods/ddb567d7-d6a7-4c5d-a54a-515098f2cce5/volumes" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.437758 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" event={"ID":"4ff57c21-f26d-4456-b388-e42583659a39","Type":"ContainerStarted","Data":"366327187acd3a992ff0dfc6cf56f86bf773471681a97f6d51be8bfe969f45b1"} Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.438168 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.440028 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b030658d-1eee-4826-b953-10f7ef16e401","Type":"ContainerStarted","Data":"b14c029a882d7a624aa1c2b0624a2f79c51ca5a45941051fa0baea9e80566862"} Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.443559 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.458405 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" podStartSLOduration=4.458388278 podStartE2EDuration="4.458388278s" podCreationTimestamp="2026-02-21 00:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:19.454830214 +0000 UTC m=+211.745665944" watchObservedRunningTime="2026-02-21 00:09:19.458388278 +0000 UTC m=+211.749224008" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.486746 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.486726581 podStartE2EDuration="2.486726581s" podCreationTimestamp="2026-02-21 00:09:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:19.485693931 +0000 UTC m=+211.776529661" watchObservedRunningTime="2026-02-21 00:09:19.486726581 +0000 UTC m=+211.777562301" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.980943 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.981836 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.984581 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.985082 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.985213 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.985312 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.985432 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.988032 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.992623 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:19 crc kubenswrapper[4730]: I0221 00:09:19.997931 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.149023 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.149372 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlq9s\" (UniqueName: \"kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.149501 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.149599 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.149678 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.251555 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.252028 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlq9s\" (UniqueName: \"kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.252116 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.252142 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.252166 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.253580 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.253723 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.254313 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.259244 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.271789 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.271884 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.279071 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlq9s\" (UniqueName: \"kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s\") pod \"controller-manager-6c5bd99cdc-ql5ff\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.297447 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.297780 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.297821 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.299078 4730 patch_prober.go:28] interesting pod/downloads-7954f5f757-bk7rn container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.299125 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-bk7rn" podUID="0fbf3e59-6719-4583-a0aa-4badc34ecf3b" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.447991 4730 generic.go:334] "Generic (PLEG): container finished" podID="b030658d-1eee-4826-b953-10f7ef16e401" containerID="b14c029a882d7a624aa1c2b0624a2f79c51ca5a45941051fa0baea9e80566862" exitCode=0 Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.448480 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b030658d-1eee-4826-b953-10f7ef16e401","Type":"ContainerDied","Data":"b14c029a882d7a624aa1c2b0624a2f79c51ca5a45941051fa0baea9e80566862"} Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.557739 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:20 crc kubenswrapper[4730]: W0221 00:09:20.569802 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbca278e0_65c3_445c_b6e3_d228fecfc714.slice/crio-2f60a006a25a613aa70cc8898aa064687cea8713fc84f7b2c62d29f12bd836f6 WatchSource:0}: Error finding container 2f60a006a25a613aa70cc8898aa064687cea8713fc84f7b2c62d29f12bd836f6: Status 404 returned error can't find the container with id 2f60a006a25a613aa70cc8898aa064687cea8713fc84f7b2c62d29f12bd836f6 Feb 21 00:09:20 crc kubenswrapper[4730]: I0221 00:09:20.858729 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.457530 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" event={"ID":"bca278e0-65c3-445c-b6e3-d228fecfc714","Type":"ContainerStarted","Data":"445661a19c3a41fe88f4e2360f9298f8199de88ca6119ee22f4761b48618cb12"} Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.457601 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" event={"ID":"bca278e0-65c3-445c-b6e3-d228fecfc714","Type":"ContainerStarted","Data":"2f60a006a25a613aa70cc8898aa064687cea8713fc84f7b2c62d29f12bd836f6"} Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.491473 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" podStartSLOduration=6.49145517 podStartE2EDuration="6.49145517s" podCreationTimestamp="2026-02-21 00:09:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:21.478461532 +0000 UTC m=+213.769297302" watchObservedRunningTime="2026-02-21 00:09:21.49145517 +0000 UTC m=+213.782290900" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.539989 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.637222 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.637612 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.713173 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.771226 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.892570 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access\") pod \"b030658d-1eee-4826-b953-10f7ef16e401\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.892663 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir\") pod \"b030658d-1eee-4826-b953-10f7ef16e401\" (UID: \"b030658d-1eee-4826-b953-10f7ef16e401\") " Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.892925 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "b030658d-1eee-4826-b953-10f7ef16e401" (UID: "b030658d-1eee-4826-b953-10f7ef16e401"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.892989 4730 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/b030658d-1eee-4826-b953-10f7ef16e401-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.903733 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "b030658d-1eee-4826-b953-10f7ef16e401" (UID: "b030658d-1eee-4826-b953-10f7ef16e401"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:21 crc kubenswrapper[4730]: I0221 00:09:21.994340 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b030658d-1eee-4826-b953-10f7ef16e401-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.046580 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.046636 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.111928 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.484117 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"b030658d-1eee-4826-b953-10f7ef16e401","Type":"ContainerDied","Data":"aa2b94bb5122a7e71df5d01c71be3262214c880ab0338a9e59ee56c1399786b7"} Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.484183 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aa2b94bb5122a7e71df5d01c71be3262214c880ab0338a9e59ee56c1399786b7" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.484436 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.486944 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.494319 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.561921 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 21 00:09:22 crc kubenswrapper[4730]: E0221 00:09:22.562315 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b030658d-1eee-4826-b953-10f7ef16e401" containerName="pruner" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.562333 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="b030658d-1eee-4826-b953-10f7ef16e401" containerName="pruner" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.562482 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="b030658d-1eee-4826-b953-10f7ef16e401" containerName="pruner" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.563139 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.563404 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.566841 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.567146 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.573408 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.574807 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.603563 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.603864 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.603960 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.704887 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.704956 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.704976 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.705091 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.705374 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.724464 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access\") pod \"installer-9-crc\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:22 crc kubenswrapper[4730]: I0221 00:09:22.881269 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:09:23 crc kubenswrapper[4730]: I0221 00:09:23.103584 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:09:23 crc kubenswrapper[4730]: I0221 00:09:23.291566 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Feb 21 00:09:23 crc kubenswrapper[4730]: W0221 00:09:23.295110 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod4b903623_cae8_4c9e_ab8f_4fb507a3a320.slice/crio-df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6 WatchSource:0}: Error finding container df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6: Status 404 returned error can't find the container with id df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6 Feb 21 00:09:23 crc kubenswrapper[4730]: I0221 00:09:23.488840 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b903623-cae8-4c9e-ab8f-4fb507a3a320","Type":"ContainerStarted","Data":"df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6"} Feb 21 00:09:24 crc kubenswrapper[4730]: I0221 00:09:24.496710 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b903623-cae8-4c9e-ab8f-4fb507a3a320","Type":"ContainerStarted","Data":"0d274ea26e176769dbb4390a675e497393422999639dade2109695c19b5de244"} Feb 21 00:09:24 crc kubenswrapper[4730]: I0221 00:09:24.497384 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bmlw" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="registry-server" containerID="cri-o://f0b3831ab0b1167b7f2b6721b728e79e167c63cc1825444be22b8327f03f16a1" gracePeriod=2 Feb 21 00:09:24 crc kubenswrapper[4730]: I0221 00:09:24.516339 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.5163204500000003 podStartE2EDuration="2.51632045s" podCreationTimestamp="2026-02-21 00:09:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:24.511763727 +0000 UTC m=+216.802599467" watchObservedRunningTime="2026-02-21 00:09:24.51632045 +0000 UTC m=+216.807156180" Feb 21 00:09:25 crc kubenswrapper[4730]: I0221 00:09:25.504652 4730 generic.go:334] "Generic (PLEG): container finished" podID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerID="f0b3831ab0b1167b7f2b6721b728e79e167c63cc1825444be22b8327f03f16a1" exitCode=0 Feb 21 00:09:25 crc kubenswrapper[4730]: I0221 00:09:25.505342 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerDied","Data":"f0b3831ab0b1167b7f2b6721b728e79e167c63cc1825444be22b8327f03f16a1"} Feb 21 00:09:26 crc kubenswrapper[4730]: I0221 00:09:26.510256 4730 generic.go:334] "Generic (PLEG): container finished" podID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerID="722cd834c8374b2df99ead2097548aeae6ec1c9c3ab27b8bd84363ec59bbfdfd" exitCode=0 Feb 21 00:09:26 crc kubenswrapper[4730]: I0221 00:09:26.510621 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerDied","Data":"722cd834c8374b2df99ead2097548aeae6ec1c9c3ab27b8bd84363ec59bbfdfd"} Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.479284 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.486924 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wlng\" (UniqueName: \"kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng\") pod \"336d55d3-841c-4b6c-9500-1405a28ac9c0\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.487000 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities\") pod \"336d55d3-841c-4b6c-9500-1405a28ac9c0\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.487043 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content\") pod \"336d55d3-841c-4b6c-9500-1405a28ac9c0\" (UID: \"336d55d3-841c-4b6c-9500-1405a28ac9c0\") " Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.488324 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities" (OuterVolumeSpecName: "utilities") pod "336d55d3-841c-4b6c-9500-1405a28ac9c0" (UID: "336d55d3-841c-4b6c-9500-1405a28ac9c0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.502610 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng" (OuterVolumeSpecName: "kube-api-access-7wlng") pod "336d55d3-841c-4b6c-9500-1405a28ac9c0" (UID: "336d55d3-841c-4b6c-9500-1405a28ac9c0"). InnerVolumeSpecName "kube-api-access-7wlng". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.524323 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bmlw" event={"ID":"336d55d3-841c-4b6c-9500-1405a28ac9c0","Type":"ContainerDied","Data":"24d75dad387611f0daedb58d6ca34943414bb4509e8bd50a5af52baaffd6718c"} Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.524618 4730 scope.go:117] "RemoveContainer" containerID="f0b3831ab0b1167b7f2b6721b728e79e167c63cc1825444be22b8327f03f16a1" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.524855 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bmlw" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.588630 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.588669 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wlng\" (UniqueName: \"kubernetes.io/projected/336d55d3-841c-4b6c-9500-1405a28ac9c0-kube-api-access-7wlng\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.819790 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "336d55d3-841c-4b6c-9500-1405a28ac9c0" (UID: "336d55d3-841c-4b6c-9500-1405a28ac9c0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.833838 4730 scope.go:117] "RemoveContainer" containerID="d7ed0e0b8e21b25cece0a7f7e047cc5aa423619f2090c38ff80cf067c283736c" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.852739 4730 scope.go:117] "RemoveContainer" containerID="42dca4c184df218d49bcc40db386e42ed66a3d0f2d83223200835e289cec8206" Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.880625 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.884963 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bmlw"] Feb 21 00:09:28 crc kubenswrapper[4730]: I0221 00:09:28.892380 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/336d55d3-841c-4b6c-9500-1405a28ac9c0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:29 crc kubenswrapper[4730]: I0221 00:09:29.531140 4730 generic.go:334] "Generic (PLEG): container finished" podID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerID="1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134" exitCode=0 Feb 21 00:09:29 crc kubenswrapper[4730]: I0221 00:09:29.531204 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerDied","Data":"1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134"} Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.320041 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-bk7rn" Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.445246 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" path="/var/lib/kubelet/pods/336d55d3-841c-4b6c-9500-1405a28ac9c0/volumes" Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.541810 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerStarted","Data":"6684f62431084975e51e98b96dcb126e0ec417b34a0b9a87e51caeaf61ee70bc"} Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.545070 4730 generic.go:334] "Generic (PLEG): container finished" podID="17855419-b9b5-4c88-a004-694eb8320ea7" containerID="d77998f1f3e84c55d274bf7d4ab13c2ac4787cac5e7d55ba4a92f8d1516cdf27" exitCode=0 Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.545155 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerDied","Data":"d77998f1f3e84c55d274bf7d4ab13c2ac4787cac5e7d55ba4a92f8d1516cdf27"} Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.548859 4730 generic.go:334] "Generic (PLEG): container finished" podID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerID="3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979" exitCode=0 Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.548988 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerDied","Data":"3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979"} Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.552598 4730 generic.go:334] "Generic (PLEG): container finished" podID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerID="ab6d5ba50a7611477525bd9fc7feabc5770b79f7d4545b84878a26cd61308f5c" exitCode=0 Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.552643 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerDied","Data":"ab6d5ba50a7611477525bd9fc7feabc5770b79f7d4545b84878a26cd61308f5c"} Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.567608 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-znbtz" podStartSLOduration=2.918651926 podStartE2EDuration="50.567586894s" podCreationTimestamp="2026-02-21 00:08:40 +0000 UTC" firstStartedPulling="2026-02-21 00:08:42.065108094 +0000 UTC m=+174.355943824" lastFinishedPulling="2026-02-21 00:09:29.714043062 +0000 UTC m=+222.004878792" observedRunningTime="2026-02-21 00:09:30.566576105 +0000 UTC m=+222.857411875" watchObservedRunningTime="2026-02-21 00:09:30.567586894 +0000 UTC m=+222.858422624" Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.654629 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:30 crc kubenswrapper[4730]: I0221 00:09:30.654686 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.561184 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerStarted","Data":"411493cf2a6ac06ae122658ad3f190a319226de5fa653da34272035f992d675d"} Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.564134 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerStarted","Data":"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82"} Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.566878 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerStarted","Data":"c044e9003fbcd11ab085b351a35b65c6bd2f9566ef07b619d3fe0cde32f02229"} Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.568924 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerStarted","Data":"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd"} Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.590478 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rk2d5" podStartSLOduration=2.298912656 podStartE2EDuration="53.590455555s" podCreationTimestamp="2026-02-21 00:08:38 +0000 UTC" firstStartedPulling="2026-02-21 00:08:40.013102327 +0000 UTC m=+172.303938057" lastFinishedPulling="2026-02-21 00:09:31.304645226 +0000 UTC m=+223.595480956" observedRunningTime="2026-02-21 00:09:31.584854442 +0000 UTC m=+223.875690172" watchObservedRunningTime="2026-02-21 00:09:31.590455555 +0000 UTC m=+223.881291315" Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.638205 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pjxx6" podStartSLOduration=3.116821636 podStartE2EDuration="53.638185251s" podCreationTimestamp="2026-02-21 00:08:38 +0000 UTC" firstStartedPulling="2026-02-21 00:08:40.018721701 +0000 UTC m=+172.309557431" lastFinishedPulling="2026-02-21 00:09:30.540085306 +0000 UTC m=+222.830921046" observedRunningTime="2026-02-21 00:09:31.635015359 +0000 UTC m=+223.925851099" watchObservedRunningTime="2026-02-21 00:09:31.638185251 +0000 UTC m=+223.929020991" Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.639390 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-d4xfg" podStartSLOduration=3.627695168 podStartE2EDuration="54.639381536s" podCreationTimestamp="2026-02-21 00:08:37 +0000 UTC" firstStartedPulling="2026-02-21 00:08:39.984074589 +0000 UTC m=+172.274910319" lastFinishedPulling="2026-02-21 00:09:30.995760957 +0000 UTC m=+223.286596687" observedRunningTime="2026-02-21 00:09:31.619063266 +0000 UTC m=+223.909899006" watchObservedRunningTime="2026-02-21 00:09:31.639381536 +0000 UTC m=+223.930217276" Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.651930 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8jnz2" podStartSLOduration=2.692522302 podStartE2EDuration="53.651897729s" podCreationTimestamp="2026-02-21 00:08:38 +0000 UTC" firstStartedPulling="2026-02-21 00:08:39.993175735 +0000 UTC m=+172.284011465" lastFinishedPulling="2026-02-21 00:09:30.952551162 +0000 UTC m=+223.243386892" observedRunningTime="2026-02-21 00:09:31.651624991 +0000 UTC m=+223.942460741" watchObservedRunningTime="2026-02-21 00:09:31.651897729 +0000 UTC m=+223.942733459" Feb 21 00:09:31 crc kubenswrapper[4730]: I0221 00:09:31.692697 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-znbtz" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="registry-server" probeResult="failure" output=< Feb 21 00:09:31 crc kubenswrapper[4730]: timeout: failed to connect service ":50051" within 1s Feb 21 00:09:31 crc kubenswrapper[4730]: > Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.211273 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.211689 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" podUID="bca278e0-65c3-445c-b6e3-d228fecfc714" containerName="controller-manager" containerID="cri-o://445661a19c3a41fe88f4e2360f9298f8199de88ca6119ee22f4761b48618cb12" gracePeriod=30 Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.227408 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.227847 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" podUID="4ff57c21-f26d-4456-b388-e42583659a39" containerName="route-controller-manager" containerID="cri-o://366327187acd3a992ff0dfc6cf56f86bf773471681a97f6d51be8bfe969f45b1" gracePeriod=30 Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.590962 4730 generic.go:334] "Generic (PLEG): container finished" podID="bca278e0-65c3-445c-b6e3-d228fecfc714" containerID="445661a19c3a41fe88f4e2360f9298f8199de88ca6119ee22f4761b48618cb12" exitCode=0 Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.591028 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" event={"ID":"bca278e0-65c3-445c-b6e3-d228fecfc714","Type":"ContainerDied","Data":"445661a19c3a41fe88f4e2360f9298f8199de88ca6119ee22f4761b48618cb12"} Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.592714 4730 generic.go:334] "Generic (PLEG): container finished" podID="4ff57c21-f26d-4456-b388-e42583659a39" containerID="366327187acd3a992ff0dfc6cf56f86bf773471681a97f6d51be8bfe969f45b1" exitCode=0 Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.592757 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" event={"ID":"4ff57c21-f26d-4456-b388-e42583659a39","Type":"ContainerDied","Data":"366327187acd3a992ff0dfc6cf56f86bf773471681a97f6d51be8bfe969f45b1"} Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.775183 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.775244 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.775288 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.775873 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c"} pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 21 00:09:35 crc kubenswrapper[4730]: I0221 00:09:35.775951 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" containerID="cri-o://9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c" gracePeriod=600 Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.601736 4730 generic.go:334] "Generic (PLEG): container finished" podID="33513de3-5480-4aef-87ff-879f9e7a475f" containerID="9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c" exitCode=0 Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.601779 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerDied","Data":"9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c"} Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.603429 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" event={"ID":"4ff57c21-f26d-4456-b388-e42583659a39","Type":"ContainerDied","Data":"906fad8c096e4291d3f258d89b566a0f026009745bf079fab8bff1632d88b5b2"} Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.603454 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="906fad8c096e4291d3f258d89b566a0f026009745bf079fab8bff1632d88b5b2" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.648116 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.686813 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:36 crc kubenswrapper[4730]: E0221 00:09:36.687051 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="extract-utilities" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687064 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="extract-utilities" Feb 21 00:09:36 crc kubenswrapper[4730]: E0221 00:09:36.687079 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="registry-server" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687086 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="registry-server" Feb 21 00:09:36 crc kubenswrapper[4730]: E0221 00:09:36.687097 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="extract-content" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687103 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="extract-content" Feb 21 00:09:36 crc kubenswrapper[4730]: E0221 00:09:36.687113 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ff57c21-f26d-4456-b388-e42583659a39" containerName="route-controller-manager" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687119 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ff57c21-f26d-4456-b388-e42583659a39" containerName="route-controller-manager" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687219 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ff57c21-f26d-4456-b388-e42583659a39" containerName="route-controller-manager" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687238 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="336d55d3-841c-4b6c-9500-1405a28ac9c0" containerName="registry-server" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.687625 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.691969 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nppsp\" (UniqueName: \"kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.692031 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.692062 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.692081 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.696465 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.730857 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.792674 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert\") pod \"4ff57c21-f26d-4456-b388-e42583659a39\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793063 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert\") pod \"bca278e0-65c3-445c-b6e3-d228fecfc714\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793106 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles\") pod \"bca278e0-65c3-445c-b6e3-d228fecfc714\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793147 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config\") pod \"bca278e0-65c3-445c-b6e3-d228fecfc714\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793172 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config\") pod \"4ff57c21-f26d-4456-b388-e42583659a39\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793219 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw6x2\" (UniqueName: \"kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2\") pod \"4ff57c21-f26d-4456-b388-e42583659a39\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793239 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlq9s\" (UniqueName: \"kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s\") pod \"bca278e0-65c3-445c-b6e3-d228fecfc714\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793263 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca\") pod \"4ff57c21-f26d-4456-b388-e42583659a39\" (UID: \"4ff57c21-f26d-4456-b388-e42583659a39\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793293 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca\") pod \"bca278e0-65c3-445c-b6e3-d228fecfc714\" (UID: \"bca278e0-65c3-445c-b6e3-d228fecfc714\") " Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793361 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793448 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nppsp\" (UniqueName: \"kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793502 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.793533 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.794818 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config" (OuterVolumeSpecName: "config") pod "4ff57c21-f26d-4456-b388-e42583659a39" (UID: "4ff57c21-f26d-4456-b388-e42583659a39"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.795065 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca" (OuterVolumeSpecName: "client-ca") pod "4ff57c21-f26d-4456-b388-e42583659a39" (UID: "4ff57c21-f26d-4456-b388-e42583659a39"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.796211 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "bca278e0-65c3-445c-b6e3-d228fecfc714" (UID: "bca278e0-65c3-445c-b6e3-d228fecfc714"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.796309 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.796541 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config" (OuterVolumeSpecName: "config") pod "bca278e0-65c3-445c-b6e3-d228fecfc714" (UID: "bca278e0-65c3-445c-b6e3-d228fecfc714"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.798682 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bca278e0-65c3-445c-b6e3-d228fecfc714" (UID: "bca278e0-65c3-445c-b6e3-d228fecfc714"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.799021 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4ff57c21-f26d-4456-b388-e42583659a39" (UID: "4ff57c21-f26d-4456-b388-e42583659a39"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.799609 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2" (OuterVolumeSpecName: "kube-api-access-cw6x2") pod "4ff57c21-f26d-4456-b388-e42583659a39" (UID: "4ff57c21-f26d-4456-b388-e42583659a39"). InnerVolumeSpecName "kube-api-access-cw6x2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.800132 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s" (OuterVolumeSpecName: "kube-api-access-xlq9s") pod "bca278e0-65c3-445c-b6e3-d228fecfc714" (UID: "bca278e0-65c3-445c-b6e3-d228fecfc714"). InnerVolumeSpecName "kube-api-access-xlq9s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.800147 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca" (OuterVolumeSpecName: "client-ca") pod "bca278e0-65c3-445c-b6e3-d228fecfc714" (UID: "bca278e0-65c3-445c-b6e3-d228fecfc714"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.800811 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.801534 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.809867 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nppsp\" (UniqueName: \"kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp\") pod \"route-controller-manager-5dc5c77675-khhvt\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894258 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894460 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw6x2\" (UniqueName: \"kubernetes.io/projected/4ff57c21-f26d-4456-b388-e42583659a39-kube-api-access-cw6x2\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894472 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlq9s\" (UniqueName: \"kubernetes.io/projected/bca278e0-65c3-445c-b6e3-d228fecfc714-kube-api-access-xlq9s\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894481 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4ff57c21-f26d-4456-b388-e42583659a39-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894491 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894499 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4ff57c21-f26d-4456-b388-e42583659a39-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894511 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bca278e0-65c3-445c-b6e3-d228fecfc714-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894519 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:36 crc kubenswrapper[4730]: I0221 00:09:36.894527 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bca278e0-65c3-445c-b6e3-d228fecfc714-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.053066 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.469036 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.619147 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" event={"ID":"bca278e0-65c3-445c-b6e3-d228fecfc714","Type":"ContainerDied","Data":"2f60a006a25a613aa70cc8898aa064687cea8713fc84f7b2c62d29f12bd836f6"} Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.619200 4730 scope.go:117] "RemoveContainer" containerID="445661a19c3a41fe88f4e2360f9298f8199de88ca6119ee22f4761b48618cb12" Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.619156 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff" Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.620894 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" event={"ID":"b5908b9b-c367-4391-975f-c1c05627a936","Type":"ContainerStarted","Data":"19263b3995715a8cf91669b28a17ccc3e0acead5613f06356cdca4530e5d7795"} Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.623545 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst" Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.626191 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587"} Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.665256 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.672122 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6c5bd99cdc-ql5ff"] Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.675683 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:37 crc kubenswrapper[4730]: I0221 00:09:37.679569 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5f4db4d774-hmnst"] Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.228214 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.228801 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.284330 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.442710 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ff57c21-f26d-4456-b388-e42583659a39" path="/var/lib/kubelet/pods/4ff57c21-f26d-4456-b388-e42583659a39/volumes" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.444130 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bca278e0-65c3-445c-b6e3-d228fecfc714" path="/var/lib/kubelet/pods/bca278e0-65c3-445c-b6e3-d228fecfc714/volumes" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.445836 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.445952 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.509833 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.630409 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" event={"ID":"b5908b9b-c367-4391-975f-c1c05627a936","Type":"ContainerStarted","Data":"f09ea39cccfb9e12a82d1581ec8e3f22ee6d13970a107b7014c4801c282948ee"} Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.631745 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.635065 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.635111 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.639299 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.665177 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" podStartSLOduration=3.665152366 podStartE2EDuration="3.665152366s" podCreationTimestamp="2026-02-21 00:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:38.657732081 +0000 UTC m=+230.948567821" watchObservedRunningTime="2026-02-21 00:09:38.665152366 +0000 UTC m=+230.955988116" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.699051 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.701909 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.702782 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.825559 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.825977 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.865324 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:38 crc kubenswrapper[4730]: I0221 00:09:38.994197 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:39 crc kubenswrapper[4730]: E0221 00:09:39.004983 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bca278e0-65c3-445c-b6e3-d228fecfc714" containerName="controller-manager" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.005036 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="bca278e0-65c3-445c-b6e3-d228fecfc714" containerName="controller-manager" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.005242 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="bca278e0-65c3-445c-b6e3-d228fecfc714" containerName="controller-manager" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.005620 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.005722 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.008768 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.009112 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.009254 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.009506 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.009566 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.012992 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.017296 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.064533 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.064626 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.064652 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b79vr\" (UniqueName: \"kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.064693 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.064849 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.165724 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.165786 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.165847 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.165872 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b79vr\" (UniqueName: \"kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.165942 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.167252 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.168473 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.168634 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.178437 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.183863 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b79vr\" (UniqueName: \"kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr\") pod \"controller-manager-5d54ff4ccb-cvpkm\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.327437 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.518325 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.638661 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" event={"ID":"2d1718bc-9e33-485d-a206-697cc48a7969","Type":"ContainerStarted","Data":"f8bfc6712dacd6cceebd01f8858f78cca0e11926e31d97a4612ed95665745bd3"} Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.687365 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:39 crc kubenswrapper[4730]: I0221 00:09:39.689542 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:40 crc kubenswrapper[4730]: I0221 00:09:40.644687 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" event={"ID":"2d1718bc-9e33-485d-a206-697cc48a7969","Type":"ContainerStarted","Data":"2dde463ae1aa15c20b7da812e8935949e02bdb409e57b1e4201f0db1a71f2478"} Feb 21 00:09:40 crc kubenswrapper[4730]: I0221 00:09:40.667080 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" podStartSLOduration=5.667065364 podStartE2EDuration="5.667065364s" podCreationTimestamp="2026-02-21 00:09:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:40.665055956 +0000 UTC m=+232.955891686" watchObservedRunningTime="2026-02-21 00:09:40.667065364 +0000 UTC m=+232.957901094" Feb 21 00:09:40 crc kubenswrapper[4730]: I0221 00:09:40.708999 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:40 crc kubenswrapper[4730]: I0221 00:09:40.748749 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:41 crc kubenswrapper[4730]: I0221 00:09:41.115020 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:09:41 crc kubenswrapper[4730]: I0221 00:09:41.660269 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-pjxx6" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="registry-server" containerID="cri-o://40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82" gracePeriod=2 Feb 21 00:09:41 crc kubenswrapper[4730]: I0221 00:09:41.662616 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:41 crc kubenswrapper[4730]: I0221 00:09:41.670098 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.060224 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.112113 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.241741 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content\") pod \"0f69aa77-802a-47fe-8367-06a3cce5a134\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.241835 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities\") pod \"0f69aa77-802a-47fe-8367-06a3cce5a134\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.241875 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l9qmc\" (UniqueName: \"kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc\") pod \"0f69aa77-802a-47fe-8367-06a3cce5a134\" (UID: \"0f69aa77-802a-47fe-8367-06a3cce5a134\") " Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.244210 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities" (OuterVolumeSpecName: "utilities") pod "0f69aa77-802a-47fe-8367-06a3cce5a134" (UID: "0f69aa77-802a-47fe-8367-06a3cce5a134"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.250289 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc" (OuterVolumeSpecName: "kube-api-access-l9qmc") pod "0f69aa77-802a-47fe-8367-06a3cce5a134" (UID: "0f69aa77-802a-47fe-8367-06a3cce5a134"). InnerVolumeSpecName "kube-api-access-l9qmc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.293861 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f69aa77-802a-47fe-8367-06a3cce5a134" (UID: "0f69aa77-802a-47fe-8367-06a3cce5a134"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.343712 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.343745 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l9qmc\" (UniqueName: \"kubernetes.io/projected/0f69aa77-802a-47fe-8367-06a3cce5a134-kube-api-access-l9qmc\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.343756 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f69aa77-802a-47fe-8367-06a3cce5a134-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.669685 4730 generic.go:334] "Generic (PLEG): container finished" podID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerID="40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82" exitCode=0 Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.669758 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerDied","Data":"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82"} Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.669968 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pjxx6" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.670187 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pjxx6" event={"ID":"0f69aa77-802a-47fe-8367-06a3cce5a134","Type":"ContainerDied","Data":"b4a3978fb98e0fde68023574a0e02c34e8b039493443797cd783096f76597f45"} Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.670249 4730 scope.go:117] "RemoveContainer" containerID="40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.670840 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8jnz2" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="registry-server" containerID="cri-o://85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd" gracePeriod=2 Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.687423 4730 scope.go:117] "RemoveContainer" containerID="1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.694414 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.697063 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-pjxx6"] Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.711973 4730 scope.go:117] "RemoveContainer" containerID="214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.747127 4730 scope.go:117] "RemoveContainer" containerID="40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82" Feb 21 00:09:42 crc kubenswrapper[4730]: E0221 00:09:42.747612 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82\": container with ID starting with 40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82 not found: ID does not exist" containerID="40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.747702 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82"} err="failed to get container status \"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82\": rpc error: code = NotFound desc = could not find container \"40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82\": container with ID starting with 40a85d6e8bdc7264672f00b3879f843027b7145f433a32b98856302006476d82 not found: ID does not exist" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.747747 4730 scope.go:117] "RemoveContainer" containerID="1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134" Feb 21 00:09:42 crc kubenswrapper[4730]: E0221 00:09:42.748385 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134\": container with ID starting with 1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134 not found: ID does not exist" containerID="1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.748422 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134"} err="failed to get container status \"1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134\": rpc error: code = NotFound desc = could not find container \"1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134\": container with ID starting with 1c626d932f17ac66925bb2eb2eaaca77bee893b46db86ccb6129d4a08cd23134 not found: ID does not exist" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.748450 4730 scope.go:117] "RemoveContainer" containerID="214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168" Feb 21 00:09:42 crc kubenswrapper[4730]: E0221 00:09:42.748784 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168\": container with ID starting with 214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168 not found: ID does not exist" containerID="214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168" Feb 21 00:09:42 crc kubenswrapper[4730]: I0221 00:09:42.748817 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168"} err="failed to get container status \"214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168\": rpc error: code = NotFound desc = could not find container \"214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168\": container with ID starting with 214ecbdc3d16526368d241f3b88ac387654d4598bdd1c4bd7fc2536a4b12d168 not found: ID does not exist" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.121534 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.260433 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content\") pod \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.260517 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6qlhc\" (UniqueName: \"kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc\") pod \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.260591 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities\") pod \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\" (UID: \"5fec200f-1602-4e98-a86a-9f1d8d633ccb\") " Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.261783 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities" (OuterVolumeSpecName: "utilities") pod "5fec200f-1602-4e98-a86a-9f1d8d633ccb" (UID: "5fec200f-1602-4e98-a86a-9f1d8d633ccb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.265336 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc" (OuterVolumeSpecName: "kube-api-access-6qlhc") pod "5fec200f-1602-4e98-a86a-9f1d8d633ccb" (UID: "5fec200f-1602-4e98-a86a-9f1d8d633ccb"). InnerVolumeSpecName "kube-api-access-6qlhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.313830 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fec200f-1602-4e98-a86a-9f1d8d633ccb" (UID: "5fec200f-1602-4e98-a86a-9f1d8d633ccb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.362621 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.362668 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6qlhc\" (UniqueName: \"kubernetes.io/projected/5fec200f-1602-4e98-a86a-9f1d8d633ccb-kube-api-access-6qlhc\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.362686 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fec200f-1602-4e98-a86a-9f1d8d633ccb-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.512754 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.513325 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-znbtz" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="registry-server" containerID="cri-o://6684f62431084975e51e98b96dcb126e0ec417b34a0b9a87e51caeaf61ee70bc" gracePeriod=2 Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.679797 4730 generic.go:334] "Generic (PLEG): container finished" podID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerID="6684f62431084975e51e98b96dcb126e0ec417b34a0b9a87e51caeaf61ee70bc" exitCode=0 Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.679861 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerDied","Data":"6684f62431084975e51e98b96dcb126e0ec417b34a0b9a87e51caeaf61ee70bc"} Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.682780 4730 generic.go:334] "Generic (PLEG): container finished" podID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerID="85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd" exitCode=0 Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.682833 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerDied","Data":"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd"} Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.682863 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8jnz2" event={"ID":"5fec200f-1602-4e98-a86a-9f1d8d633ccb","Type":"ContainerDied","Data":"bc7bcc1e3812adcc0ac62d7802930d1d9f6e6908426c81ba11cee27ae65b4e00"} Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.682890 4730 scope.go:117] "RemoveContainer" containerID="85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.682944 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8jnz2" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.735003 4730 scope.go:117] "RemoveContainer" containerID="3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.745464 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.748524 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8jnz2"] Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.816635 4730 scope.go:117] "RemoveContainer" containerID="fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.864662 4730 scope.go:117] "RemoveContainer" containerID="85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd" Feb 21 00:09:43 crc kubenswrapper[4730]: E0221 00:09:43.868643 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd\": container with ID starting with 85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd not found: ID does not exist" containerID="85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.868688 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd"} err="failed to get container status \"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd\": rpc error: code = NotFound desc = could not find container \"85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd\": container with ID starting with 85afa1f788f6fb1b342c6b8ac4ea9a31c1f0c290e388a4ed91d9e9c19665bdcd not found: ID does not exist" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.868719 4730 scope.go:117] "RemoveContainer" containerID="3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979" Feb 21 00:09:43 crc kubenswrapper[4730]: E0221 00:09:43.870413 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979\": container with ID starting with 3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979 not found: ID does not exist" containerID="3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.870482 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979"} err="failed to get container status \"3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979\": rpc error: code = NotFound desc = could not find container \"3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979\": container with ID starting with 3db0ab7fd55a163726d21df7ef8c4f477ecc870e1e38b7e605c535d44aa99979 not found: ID does not exist" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.870546 4730 scope.go:117] "RemoveContainer" containerID="fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd" Feb 21 00:09:43 crc kubenswrapper[4730]: E0221 00:09:43.871466 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd\": container with ID starting with fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd not found: ID does not exist" containerID="fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd" Feb 21 00:09:43 crc kubenswrapper[4730]: I0221 00:09:43.871492 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd"} err="failed to get container status \"fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd\": rpc error: code = NotFound desc = could not find container \"fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd\": container with ID starting with fca3d45571f9402eff79840d33968cfb4fbb874f936f7573e03a5722c33945fd not found: ID does not exist" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.005333 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.195273 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content\") pod \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.195356 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdlhw\" (UniqueName: \"kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw\") pod \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.195393 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities\") pod \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\" (UID: \"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb\") " Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.196461 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities" (OuterVolumeSpecName: "utilities") pod "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" (UID: "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.200414 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw" (OuterVolumeSpecName: "kube-api-access-gdlhw") pod "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" (UID: "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb"). InnerVolumeSpecName "kube-api-access-gdlhw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.218962 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" (UID: "e68cc82a-3bb4-40df-bab5-a9b4fa735bbb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.296237 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdlhw\" (UniqueName: \"kubernetes.io/projected/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-kube-api-access-gdlhw\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.296265 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.296275 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.445187 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" path="/var/lib/kubelet/pods/0f69aa77-802a-47fe-8367-06a3cce5a134/volumes" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.447452 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" path="/var/lib/kubelet/pods/5fec200f-1602-4e98-a86a-9f1d8d633ccb/volumes" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.715464 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-znbtz" event={"ID":"e68cc82a-3bb4-40df-bab5-a9b4fa735bbb","Type":"ContainerDied","Data":"5e8590b10d4fae4e4284ac206eda32330bdaab27f522820584e6d9c92bc37d99"} Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.715524 4730 scope.go:117] "RemoveContainer" containerID="6684f62431084975e51e98b96dcb126e0ec417b34a0b9a87e51caeaf61ee70bc" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.715561 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-znbtz" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.737597 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.738377 4730 scope.go:117] "RemoveContainer" containerID="722cd834c8374b2df99ead2097548aeae6ec1c9c3ab27b8bd84363ec59bbfdfd" Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.741977 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-znbtz"] Feb 21 00:09:44 crc kubenswrapper[4730]: I0221 00:09:44.753227 4730 scope.go:117] "RemoveContainer" containerID="aa068a1421584a957dfd93624e21bf4c9adcccfa068353cfbc124a3ada16d734" Feb 21 00:09:46 crc kubenswrapper[4730]: I0221 00:09:46.446459 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" path="/var/lib/kubelet/pods/e68cc82a-3bb4-40df-bab5-a9b4fa735bbb/volumes" Feb 21 00:09:48 crc kubenswrapper[4730]: I0221 00:09:48.389767 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ssd6c"] Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.219290 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.220175 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" podUID="2d1718bc-9e33-485d-a206-697cc48a7969" containerName="controller-manager" containerID="cri-o://2dde463ae1aa15c20b7da812e8935949e02bdb409e57b1e4201f0db1a71f2478" gracePeriod=30 Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.317652 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.318000 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" podUID="b5908b9b-c367-4391-975f-c1c05627a936" containerName="route-controller-manager" containerID="cri-o://f09ea39cccfb9e12a82d1581ec8e3f22ee6d13970a107b7014c4801c282948ee" gracePeriod=30 Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.778118 4730 generic.go:334] "Generic (PLEG): container finished" podID="b5908b9b-c367-4391-975f-c1c05627a936" containerID="f09ea39cccfb9e12a82d1581ec8e3f22ee6d13970a107b7014c4801c282948ee" exitCode=0 Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.778230 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" event={"ID":"b5908b9b-c367-4391-975f-c1c05627a936","Type":"ContainerDied","Data":"f09ea39cccfb9e12a82d1581ec8e3f22ee6d13970a107b7014c4801c282948ee"} Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.779758 4730 generic.go:334] "Generic (PLEG): container finished" podID="2d1718bc-9e33-485d-a206-697cc48a7969" containerID="2dde463ae1aa15c20b7da812e8935949e02bdb409e57b1e4201f0db1a71f2478" exitCode=0 Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.779780 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" event={"ID":"2d1718bc-9e33-485d-a206-697cc48a7969","Type":"ContainerDied","Data":"2dde463ae1aa15c20b7da812e8935949e02bdb409e57b1e4201f0db1a71f2478"} Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.838451 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.839017 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880552 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config\") pod \"b5908b9b-c367-4391-975f-c1c05627a936\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880620 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config\") pod \"2d1718bc-9e33-485d-a206-697cc48a7969\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880673 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert\") pod \"2d1718bc-9e33-485d-a206-697cc48a7969\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880690 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles\") pod \"2d1718bc-9e33-485d-a206-697cc48a7969\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880756 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert\") pod \"b5908b9b-c367-4391-975f-c1c05627a936\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880785 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca\") pod \"2d1718bc-9e33-485d-a206-697cc48a7969\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880821 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b79vr\" (UniqueName: \"kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr\") pod \"2d1718bc-9e33-485d-a206-697cc48a7969\" (UID: \"2d1718bc-9e33-485d-a206-697cc48a7969\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880862 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca\") pod \"b5908b9b-c367-4391-975f-c1c05627a936\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.880894 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nppsp\" (UniqueName: \"kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp\") pod \"b5908b9b-c367-4391-975f-c1c05627a936\" (UID: \"b5908b9b-c367-4391-975f-c1c05627a936\") " Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.881365 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config" (OuterVolumeSpecName: "config") pod "b5908b9b-c367-4391-975f-c1c05627a936" (UID: "b5908b9b-c367-4391-975f-c1c05627a936"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.881782 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca" (OuterVolumeSpecName: "client-ca") pod "b5908b9b-c367-4391-975f-c1c05627a936" (UID: "b5908b9b-c367-4391-975f-c1c05627a936"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.882001 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "2d1718bc-9e33-485d-a206-697cc48a7969" (UID: "2d1718bc-9e33-485d-a206-697cc48a7969"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.882034 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config" (OuterVolumeSpecName: "config") pod "2d1718bc-9e33-485d-a206-697cc48a7969" (UID: "2d1718bc-9e33-485d-a206-697cc48a7969"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.882226 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca" (OuterVolumeSpecName: "client-ca") pod "2d1718bc-9e33-485d-a206-697cc48a7969" (UID: "2d1718bc-9e33-485d-a206-697cc48a7969"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.887433 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b5908b9b-c367-4391-975f-c1c05627a936" (UID: "b5908b9b-c367-4391-975f-c1c05627a936"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.888325 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr" (OuterVolumeSpecName: "kube-api-access-b79vr") pod "2d1718bc-9e33-485d-a206-697cc48a7969" (UID: "2d1718bc-9e33-485d-a206-697cc48a7969"). InnerVolumeSpecName "kube-api-access-b79vr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.888441 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp" (OuterVolumeSpecName: "kube-api-access-nppsp") pod "b5908b9b-c367-4391-975f-c1c05627a936" (UID: "b5908b9b-c367-4391-975f-c1c05627a936"). InnerVolumeSpecName "kube-api-access-nppsp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.888876 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2d1718bc-9e33-485d-a206-697cc48a7969" (UID: "2d1718bc-9e33-485d-a206-697cc48a7969"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982105 4730 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982138 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2d1718bc-9e33-485d-a206-697cc48a7969-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982147 4730 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b5908b9b-c367-4391-975f-c1c05627a936-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982155 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982163 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b79vr\" (UniqueName: \"kubernetes.io/projected/2d1718bc-9e33-485d-a206-697cc48a7969-kube-api-access-b79vr\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982173 4730 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-client-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982182 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nppsp\" (UniqueName: \"kubernetes.io/projected/b5908b9b-c367-4391-975f-c1c05627a936-kube-api-access-nppsp\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982190 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b5908b9b-c367-4391-975f-c1c05627a936-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:55 crc kubenswrapper[4730]: I0221 00:09:55.982199 4730 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2d1718bc-9e33-485d-a206-697cc48a7969-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.787408 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" event={"ID":"b5908b9b-c367-4391-975f-c1c05627a936","Type":"ContainerDied","Data":"19263b3995715a8cf91669b28a17ccc3e0acead5613f06356cdca4530e5d7795"} Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.787469 4730 scope.go:117] "RemoveContainer" containerID="f09ea39cccfb9e12a82d1581ec8e3f22ee6d13970a107b7014c4801c282948ee" Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.787465 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt" Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.789344 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" event={"ID":"2d1718bc-9e33-485d-a206-697cc48a7969","Type":"ContainerDied","Data":"f8bfc6712dacd6cceebd01f8858f78cca0e11926e31d97a4612ed95665745bd3"} Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.789372 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm" Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.808579 4730 scope.go:117] "RemoveContainer" containerID="2dde463ae1aa15c20b7da812e8935949e02bdb409e57b1e4201f0db1a71f2478" Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.824868 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.848724 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5dc5c77675-khhvt"] Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.857860 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:56 crc kubenswrapper[4730]: I0221 00:09:56.861092 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5d54ff4ccb-cvpkm"] Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.013803 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw"] Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014270 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014293 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014307 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014315 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014329 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014337 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014349 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d1718bc-9e33-485d-a206-697cc48a7969" containerName="controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014357 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d1718bc-9e33-485d-a206-697cc48a7969" containerName="controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014366 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014374 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014386 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014394 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014407 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5908b9b-c367-4391-975f-c1c05627a936" containerName="route-controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014415 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5908b9b-c367-4391-975f-c1c05627a936" containerName="route-controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014427 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014434 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="extract-content" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014448 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014456 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014467 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014474 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: E0221 00:09:57.014489 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014499 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="extract-utilities" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014610 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fec200f-1602-4e98-a86a-9f1d8d633ccb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014624 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5908b9b-c367-4391-975f-c1c05627a936" containerName="route-controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014639 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f69aa77-802a-47fe-8367-06a3cce5a134" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014649 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d1718bc-9e33-485d-a206-697cc48a7969" containerName="controller-manager" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.014658 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68cc82a-3bb4-40df-bab5-a9b4fa735bbb" containerName="registry-server" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.015174 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.017696 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.018039 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.018155 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.018231 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.018494 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.018718 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.021336 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-588c7477f7-5kbj9"] Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.022622 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.025971 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.026302 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.026831 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.027503 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.027596 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-588c7477f7-5kbj9"] Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.027627 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.027666 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.032777 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.034990 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw"] Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.095743 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-config\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.095799 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-proxy-ca-bundles\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.095856 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-config\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.095886 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kl7dl\" (UniqueName: \"kubernetes.io/projected/84ff14de-f655-4a2b-8d7a-328a2f9829fc-kube-api-access-kl7dl\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.096006 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jj742\" (UniqueName: \"kubernetes.io/projected/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-kube-api-access-jj742\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.096062 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ff14de-f655-4a2b-8d7a-328a2f9829fc-serving-cert\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.096230 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-client-ca\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.096280 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-client-ca\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.096357 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-serving-cert\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197314 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-config\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197360 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-proxy-ca-bundles\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197387 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-config\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197405 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kl7dl\" (UniqueName: \"kubernetes.io/projected/84ff14de-f655-4a2b-8d7a-328a2f9829fc-kube-api-access-kl7dl\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197428 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jj742\" (UniqueName: \"kubernetes.io/projected/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-kube-api-access-jj742\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197449 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ff14de-f655-4a2b-8d7a-328a2f9829fc-serving-cert\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197486 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-client-ca\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197501 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-client-ca\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.197543 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-serving-cert\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.199212 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-proxy-ca-bundles\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.199397 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-client-ca\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.199487 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/84ff14de-f655-4a2b-8d7a-328a2f9829fc-config\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.199510 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-config\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.200051 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-client-ca\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.202326 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-serving-cert\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.205373 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/84ff14de-f655-4a2b-8d7a-328a2f9829fc-serving-cert\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.220360 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kl7dl\" (UniqueName: \"kubernetes.io/projected/84ff14de-f655-4a2b-8d7a-328a2f9829fc-kube-api-access-kl7dl\") pod \"route-controller-manager-776d96c948-6qsmw\" (UID: \"84ff14de-f655-4a2b-8d7a-328a2f9829fc\") " pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.222274 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jj742\" (UniqueName: \"kubernetes.io/projected/bdfae9ba-2ccf-4a4d-b670-efb940a111cf-kube-api-access-jj742\") pod \"controller-manager-588c7477f7-5kbj9\" (UID: \"bdfae9ba-2ccf-4a4d-b670-efb940a111cf\") " pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.348198 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.358077 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.832813 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-588c7477f7-5kbj9"] Feb 21 00:09:57 crc kubenswrapper[4730]: W0221 00:09:57.837070 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbdfae9ba_2ccf_4a4d_b670_efb940a111cf.slice/crio-18e60332cb13f4c270194db36cbcce57d599d93a1ece45e6ebbcd6375124a6be WatchSource:0}: Error finding container 18e60332cb13f4c270194db36cbcce57d599d93a1ece45e6ebbcd6375124a6be: Status 404 returned error can't find the container with id 18e60332cb13f4c270194db36cbcce57d599d93a1ece45e6ebbcd6375124a6be Feb 21 00:09:57 crc kubenswrapper[4730]: I0221 00:09:57.839614 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw"] Feb 21 00:09:57 crc kubenswrapper[4730]: W0221 00:09:57.850857 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84ff14de_f655_4a2b_8d7a_328a2f9829fc.slice/crio-ee1097aa6dd2090ab84b2f1e5e4ed8eeccf875e3ffcf9782fd27034ba32c67ed WatchSource:0}: Error finding container ee1097aa6dd2090ab84b2f1e5e4ed8eeccf875e3ffcf9782fd27034ba32c67ed: Status 404 returned error can't find the container with id ee1097aa6dd2090ab84b2f1e5e4ed8eeccf875e3ffcf9782fd27034ba32c67ed Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.446957 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d1718bc-9e33-485d-a206-697cc48a7969" path="/var/lib/kubelet/pods/2d1718bc-9e33-485d-a206-697cc48a7969/volumes" Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.447890 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5908b9b-c367-4391-975f-c1c05627a936" path="/var/lib/kubelet/pods/b5908b9b-c367-4391-975f-c1c05627a936/volumes" Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.806651 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" event={"ID":"bdfae9ba-2ccf-4a4d-b670-efb940a111cf","Type":"ContainerStarted","Data":"1ff3861a02bb87e14b2e0be465d5336af02b438af14c8f08a2dec49cb8fdc804"} Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.806700 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" event={"ID":"bdfae9ba-2ccf-4a4d-b670-efb940a111cf","Type":"ContainerStarted","Data":"18e60332cb13f4c270194db36cbcce57d599d93a1ece45e6ebbcd6375124a6be"} Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.808849 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" event={"ID":"84ff14de-f655-4a2b-8d7a-328a2f9829fc","Type":"ContainerStarted","Data":"17656b6cf49ca4ecba7283cbb6e85739ff4dbf92092002d5eefa9021686c1923"} Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.809007 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" event={"ID":"84ff14de-f655-4a2b-8d7a-328a2f9829fc","Type":"ContainerStarted","Data":"ee1097aa6dd2090ab84b2f1e5e4ed8eeccf875e3ffcf9782fd27034ba32c67ed"} Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.809210 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.815075 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.831429 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" podStartSLOduration=3.831397625 podStartE2EDuration="3.831397625s" podCreationTimestamp="2026-02-21 00:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:58.823011161 +0000 UTC m=+251.113846891" watchObservedRunningTime="2026-02-21 00:09:58.831397625 +0000 UTC m=+251.122233395" Feb 21 00:09:58 crc kubenswrapper[4730]: I0221 00:09:58.847740 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-776d96c948-6qsmw" podStartSLOduration=3.847714729 podStartE2EDuration="3.847714729s" podCreationTimestamp="2026-02-21 00:09:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:09:58.839014866 +0000 UTC m=+251.129850616" watchObservedRunningTime="2026-02-21 00:09:58.847714729 +0000 UTC m=+251.138550489" Feb 21 00:09:59 crc kubenswrapper[4730]: I0221 00:09:59.815084 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:09:59 crc kubenswrapper[4730]: I0221 00:09:59.820399 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-588c7477f7-5kbj9" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.568119 4730 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.569009 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.572467 4730 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.572851 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40" gracePeriod=15 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.572938 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f" gracePeriod=15 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.572996 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520" gracePeriod=15 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.573070 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb" gracePeriod=15 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.573009 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad" gracePeriod=15 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.576749 4730 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577109 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577131 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577153 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577163 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577178 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577188 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577202 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577211 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577231 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577239 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577251 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577262 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.577272 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577281 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577464 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577484 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577498 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577509 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577523 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.577535 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.639451 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674543 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674586 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674607 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674623 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674717 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674750 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674774 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.674804 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775411 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775724 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775749 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775776 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775790 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775794 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775804 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775511 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775835 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775847 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775853 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775871 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775872 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775830 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775927 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.775982 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.829154 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.831243 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.832191 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f" exitCode=0 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.832231 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520" exitCode=0 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.832248 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad" exitCode=0 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.832265 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb" exitCode=2 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.832319 4730 scope.go:117] "RemoveContainer" containerID="e6744fa13c88782ade24a8a6ed2535473fa98bef3c5e3e583145ee1bb1e9a6dc" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.833590 4730 generic.go:334] "Generic (PLEG): container finished" podID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" containerID="0d274ea26e176769dbb4390a675e497393422999639dade2109695c19b5de244" exitCode=0 Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.833633 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b903623-cae8-4c9e-ab8f-4fb507a3a320","Type":"ContainerDied","Data":"0d274ea26e176769dbb4390a675e497393422999639dade2109695c19b5de244"} Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.834674 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.835255 4730 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.835722 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.848368 4730 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" start-of-body= Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.848419 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": dial tcp 192.168.126.11:6443: connect: connection refused" Feb 21 00:10:01 crc kubenswrapper[4730]: I0221 00:10:01.934794 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:10:01 crc kubenswrapper[4730]: W0221 00:10:01.961653 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c55f741074fc910be05bc1e509ebf33f1a5e4aa9bbe3ec1feed30d116fd716eb WatchSource:0}: Error finding container c55f741074fc910be05bc1e509ebf33f1a5e4aa9bbe3ec1feed30d116fd716eb: Status 404 returned error can't find the container with id c55f741074fc910be05bc1e509ebf33f1a5e4aa9bbe3ec1feed30d116fd716eb Feb 21 00:10:01 crc kubenswrapper[4730]: E0221 00:10:01.967370 4730 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.179:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18961a7708b8569c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-21 00:10:01.966597788 +0000 UTC m=+254.257433558,LastTimestamp:2026-02-21 00:10:01.966597788 +0000 UTC m=+254.257433558,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 21 00:10:02 crc kubenswrapper[4730]: I0221 00:10:02.849522 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 21 00:10:02 crc kubenswrapper[4730]: I0221 00:10:02.854529 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"37c6eb74e229c8794c48181f96eb406807a6a6bccbdd1e28a91cbe38147a06eb"} Feb 21 00:10:02 crc kubenswrapper[4730]: I0221 00:10:02.854575 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c55f741074fc910be05bc1e509ebf33f1a5e4aa9bbe3ec1feed30d116fd716eb"} Feb 21 00:10:02 crc kubenswrapper[4730]: I0221 00:10:02.855314 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:02 crc kubenswrapper[4730]: I0221 00:10:02.855669 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.357591 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.358393 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.359100 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.501759 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock\") pod \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.501843 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access\") pod \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.501866 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir\") pod \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\" (UID: \"4b903623-cae8-4c9e-ab8f-4fb507a3a320\") " Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.501977 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock" (OuterVolumeSpecName: "var-lock") pod "4b903623-cae8-4c9e-ab8f-4fb507a3a320" (UID: "4b903623-cae8-4c9e-ab8f-4fb507a3a320"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.502225 4730 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-var-lock\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.502207 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "4b903623-cae8-4c9e-ab8f-4fb507a3a320" (UID: "4b903623-cae8-4c9e-ab8f-4fb507a3a320"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.511183 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "4b903623-cae8-4c9e-ab8f-4fb507a3a320" (UID: "4b903623-cae8-4c9e-ab8f-4fb507a3a320"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.604783 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kube-api-access\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.604989 4730 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/4b903623-cae8-4c9e-ab8f-4fb507a3a320-kubelet-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.861101 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"4b903623-cae8-4c9e-ab8f-4fb507a3a320","Type":"ContainerDied","Data":"df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6"} Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.861385 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df0227a445cef3dd0c97b13d504848f9c710d48e11b134b781da8ee131c951c6" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.861146 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.875326 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:03 crc kubenswrapper[4730]: I0221 00:10:03.875724 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.012690 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.013872 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.014403 4730 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.015010 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.015530 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.212987 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213121 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213183 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213255 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213300 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213426 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213634 4730 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213652 4730 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.213663 4730 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.450149 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.873882 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.874668 4730 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40" exitCode=0 Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.874732 4730 scope.go:117] "RemoveContainer" containerID="937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.874735 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.875204 4730 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.875627 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.875886 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.879113 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.879532 4730 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.879958 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.909637 4730 scope.go:117] "RemoveContainer" containerID="48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.923805 4730 scope.go:117] "RemoveContainer" containerID="3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.941336 4730 scope.go:117] "RemoveContainer" containerID="83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.955303 4730 scope.go:117] "RemoveContainer" containerID="eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.971069 4730 scope.go:117] "RemoveContainer" containerID="13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.993808 4730 scope.go:117] "RemoveContainer" containerID="937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.994197 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\": container with ID starting with 937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f not found: ID does not exist" containerID="937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.994238 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f"} err="failed to get container status \"937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\": rpc error: code = NotFound desc = could not find container \"937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f\": container with ID starting with 937a1c9af4a04f117363acb26a14c6d6c01167ab5d7380276d970dd45c08924f not found: ID does not exist" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.994267 4730 scope.go:117] "RemoveContainer" containerID="48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.996024 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\": container with ID starting with 48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520 not found: ID does not exist" containerID="48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996064 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520"} err="failed to get container status \"48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\": rpc error: code = NotFound desc = could not find container \"48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520\": container with ID starting with 48829d5593bef176c85efc223c39f30e35985097f12a06c78f20bdc07765b520 not found: ID does not exist" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996110 4730 scope.go:117] "RemoveContainer" containerID="3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.996485 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\": container with ID starting with 3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad not found: ID does not exist" containerID="3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996526 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad"} err="failed to get container status \"3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\": rpc error: code = NotFound desc = could not find container \"3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad\": container with ID starting with 3cd7e4d4bd3408452d0ab2ad5425837f6dda742ae682fdb745a0ef0f5e8e36ad not found: ID does not exist" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996553 4730 scope.go:117] "RemoveContainer" containerID="83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.996864 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\": container with ID starting with 83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb not found: ID does not exist" containerID="83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996895 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb"} err="failed to get container status \"83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\": rpc error: code = NotFound desc = could not find container \"83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb\": container with ID starting with 83500ee6e48f4218673c8a578eff0fed47a77e876ea4017cfc58bcec3daf39bb not found: ID does not exist" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.996928 4730 scope.go:117] "RemoveContainer" containerID="eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.997211 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\": container with ID starting with eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40 not found: ID does not exist" containerID="eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.997232 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40"} err="failed to get container status \"eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\": rpc error: code = NotFound desc = could not find container \"eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40\": container with ID starting with eb605d66f039d6f071f39891f3550b73f9a8ca8fa350c75673f10fdc939edd40 not found: ID does not exist" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.997247 4730 scope.go:117] "RemoveContainer" containerID="13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400" Feb 21 00:10:04 crc kubenswrapper[4730]: E0221 00:10:04.997826 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\": container with ID starting with 13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400 not found: ID does not exist" containerID="13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400" Feb 21 00:10:04 crc kubenswrapper[4730]: I0221 00:10:04.997885 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400"} err="failed to get container status \"13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\": rpc error: code = NotFound desc = could not find container \"13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400\": container with ID starting with 13d1e93e6dd79f3c7acbd735462efd293146e5015ea20efe455441cba5c4f400 not found: ID does not exist" Feb 21 00:10:05 crc kubenswrapper[4730]: E0221 00:10:05.886831 4730 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.179:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.18961a7708b8569c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-02-21 00:10:01.966597788 +0000 UTC m=+254.257433558,LastTimestamp:2026-02-21 00:10:01.966597788 +0000 UTC m=+254.257433558,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.038563 4730 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.039872 4730 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.040654 4730 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.041500 4730 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.041869 4730 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: I0221 00:10:08.041950 4730 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.042333 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="200ms" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.243841 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="400ms" Feb 21 00:10:08 crc kubenswrapper[4730]: I0221 00:10:08.439714 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: I0221 00:10:08.440451 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.572379 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:10:08Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:10:08Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:10:08Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-02-21T00:10:08Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.573439 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.574168 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.574710 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.575355 4730 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.575395 4730 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Feb 21 00:10:08 crc kubenswrapper[4730]: E0221 00:10:08.645577 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="800ms" Feb 21 00:10:09 crc kubenswrapper[4730]: E0221 00:10:09.446633 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="1.6s" Feb 21 00:10:11 crc kubenswrapper[4730]: E0221 00:10:11.048009 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="3.2s" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.435011 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.436387 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.436966 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.448714 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" containerName="oauth-openshift" containerID="cri-o://5e5cc3d04d9aa6416b05d9eed939ef8824c602fd2b4709cdea7faa531f23753a" gracePeriod=15 Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.474793 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.474860 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:13 crc kubenswrapper[4730]: E0221 00:10:13.475653 4730 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.476588 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.928722 4730 generic.go:334] "Generic (PLEG): container finished" podID="cf9c54fa-0743-40ed-85b7-ae38607f7265" containerID="5e5cc3d04d9aa6416b05d9eed939ef8824c602fd2b4709cdea7faa531f23753a" exitCode=0 Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.928880 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" event={"ID":"cf9c54fa-0743-40ed-85b7-ae38607f7265","Type":"ContainerDied","Data":"5e5cc3d04d9aa6416b05d9eed939ef8824c602fd2b4709cdea7faa531f23753a"} Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.931832 4730 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="a4f43b24fc67d108652d5edf3985de99a399a76ffcb89a8c00a2b5f723a78e74" exitCode=0 Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.931872 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"a4f43b24fc67d108652d5edf3985de99a399a76ffcb89a8c00a2b5f723a78e74"} Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.931916 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"52000104af0be7821b5961752998fe5b28921b222eeabd5ecc33b888c9b91a05"} Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.932181 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.932199 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:13 crc kubenswrapper[4730]: E0221 00:10:13.932520 4730 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.932578 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.932821 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.989004 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.989968 4730 status_manager.go:851] "Failed to get status for pod" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-ssd6c\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.990605 4730 status_manager.go:851] "Failed to get status for pod" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:13 crc kubenswrapper[4730]: I0221 00:10:13.991191 4730 status_manager.go:851] "Failed to get status for pod" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-startup-monitor-crc\": dial tcp 38.102.83.179:6443: connect: connection refused" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.065980 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066030 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ll5bt\" (UniqueName: \"kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066067 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066096 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066118 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066164 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066188 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066213 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066234 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066267 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066294 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066314 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066335 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.066359 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir\") pod \"cf9c54fa-0743-40ed-85b7-ae38607f7265\" (UID: \"cf9c54fa-0743-40ed-85b7-ae38607f7265\") " Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.067851 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.068131 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.070344 4730 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-policies\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.070374 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.070411 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.072666 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt" (OuterVolumeSpecName: "kube-api-access-ll5bt") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "kube-api-access-ll5bt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.072982 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073272 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073365 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073488 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073557 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073548 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.073720 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.076243 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.081115 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.092821 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "cf9c54fa-0743-40ed-85b7-ae38607f7265" (UID: "cf9c54fa-0743-40ed-85b7-ae38607f7265"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171630 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171680 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171696 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171710 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171725 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171737 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171749 4730 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/cf9c54fa-0743-40ed-85b7-ae38607f7265-audit-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171764 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171777 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ll5bt\" (UniqueName: \"kubernetes.io/projected/cf9c54fa-0743-40ed-85b7-ae38607f7265-kube-api-access-ll5bt\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171789 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171801 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.171813 4730 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/cf9c54fa-0743-40ed-85b7-ae38607f7265-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:14 crc kubenswrapper[4730]: E0221 00:10:14.250659 4730 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.179:6443: connect: connection refused" interval="6.4s" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.943418 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.943385 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-ssd6c" event={"ID":"cf9c54fa-0743-40ed-85b7-ae38607f7265","Type":"ContainerDied","Data":"75cbe7e5c43dfbcca2dba4494942383466d8d7a8ca697f76ced6eab437575552"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.943896 4730 scope.go:117] "RemoveContainer" containerID="5e5cc3d04d9aa6416b05d9eed939ef8824c602fd2b4709cdea7faa531f23753a" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.965219 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7cce8851e70422bfb97fc4fc06407a8383898eac535bae0c6df51120cc095df6"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.965285 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"140c21c2435c918f50eb70ef047e7bfa338f52311f1b1668e9f7c47b64dae3e6"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.965300 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"63cbe6a86245c14dbd15c947e84c11294fab5a607be455f735970042b4fa1847"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.965312 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"29d03fbeacdc8a80042228d8a65c34732efee44797c45a63806bab43fabd4fcb"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.969271 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.969313 4730 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740" exitCode=1 Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.969341 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740"} Feb 21 00:10:14 crc kubenswrapper[4730]: I0221 00:10:14.969759 4730 scope.go:117] "RemoveContainer" containerID="8dd9155f6ec906b7a04044119135cec2c4ef912347fecaac02a1b219aeb59740" Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.980271 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1d5e38eb11fa2f5d5fe26811d9b8be24aa9819513df624817490926a9bffd0d8"} Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.981738 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.982056 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.982288 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.983941 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Feb 21 00:10:15 crc kubenswrapper[4730]: I0221 00:10:15.983996 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f135a7ba4b407dd073df30829ae1b26b5745cdfe3a356c57725762b4a108b48f"} Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.443671 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.444161 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.446392 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.457311 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.474593 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.477725 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.477786 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.482988 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.499304 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.647327 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.647406 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.649717 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.651246 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.659229 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.661394 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.768583 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.777704 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Feb 21 00:10:18 crc kubenswrapper[4730]: I0221 00:10:18.781003 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:10:19 crc kubenswrapper[4730]: W0221 00:10:19.343496 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-5e54279bb0979086280b3fc4630ae18c820b30d388e4ef69ad7a781bc0569bc2 WatchSource:0}: Error finding container 5e54279bb0979086280b3fc4630ae18c820b30d388e4ef69ad7a781bc0569bc2: Status 404 returned error can't find the container with id 5e54279bb0979086280b3fc4630ae18c820b30d388e4ef69ad7a781bc0569bc2 Feb 21 00:10:19 crc kubenswrapper[4730]: I0221 00:10:19.449064 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:10:19 crc kubenswrapper[4730]: I0221 00:10:19.455213 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:10:19 crc kubenswrapper[4730]: W0221 00:10:19.536562 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-b777d67a950ef07d32f394361c9cc3d0667f45b26ffc3f23b4b177310c0e06fe WatchSource:0}: Error finding container b777d67a950ef07d32f394361c9cc3d0667f45b26ffc3f23b4b177310c0e06fe: Status 404 returned error can't find the container with id b777d67a950ef07d32f394361c9cc3d0667f45b26ffc3f23b4b177310c0e06fe Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.010860 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"85324e0b3d0728137a23904f5bd87adb2b08b2ade3e13ba961023afee57ddf9e"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.011331 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"730d7ecbc55a55c1634e25d97c952b70015917d83e75e205576307ac0c11879a"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.014794 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"0b04be351cdfc8f0fc70a27289e27794a1e9314192ae76f240f0702f6f0a83e4"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.014845 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5e54279bb0979086280b3fc4630ae18c820b30d388e4ef69ad7a781bc0569bc2"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.018069 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"179fd528aab74b0340220c0a31d5c2c7ed914febb0dbc7b041c776b90f6828b9"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.018124 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.018153 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"b777d67a950ef07d32f394361c9cc3d0667f45b26ffc3f23b4b177310c0e06fe"} Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.018821 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:10:20 crc kubenswrapper[4730]: I0221 00:10:20.998574 4730 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.034778 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.035169 4730 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="85324e0b3d0728137a23904f5bd87adb2b08b2ade3e13ba961023afee57ddf9e" exitCode=255 Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.035316 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"85324e0b3d0728137a23904f5bd87adb2b08b2ade3e13ba961023afee57ddf9e"} Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.035635 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.035649 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.037234 4730 scope.go:117] "RemoveContainer" containerID="85324e0b3d0728137a23904f5bd87adb2b08b2ade3e13ba961023afee57ddf9e" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.042938 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:22 crc kubenswrapper[4730]: I0221 00:10:22.063734 4730 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4bd136f8-bcf2-4ddf-b885-d9455bd1b1a9" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.044334 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.046756 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/0.log" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.046884 4730 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="8b2a0ba92a6cea58b687f61b8ca716a2e782ad98635dcb1ed0cba58bcdfc87bd" exitCode=255 Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.047048 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"8b2a0ba92a6cea58b687f61b8ca716a2e782ad98635dcb1ed0cba58bcdfc87bd"} Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.047106 4730 scope.go:117] "RemoveContainer" containerID="85324e0b3d0728137a23904f5bd87adb2b08b2ade3e13ba961023afee57ddf9e" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.047441 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.047477 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:23 crc kubenswrapper[4730]: I0221 00:10:23.048147 4730 scope.go:117] "RemoveContainer" containerID="8b2a0ba92a6cea58b687f61b8ca716a2e782ad98635dcb1ed0cba58bcdfc87bd" Feb 21 00:10:23 crc kubenswrapper[4730]: E0221 00:10:23.048852 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:10:24 crc kubenswrapper[4730]: I0221 00:10:24.056671 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 21 00:10:28 crc kubenswrapper[4730]: I0221 00:10:28.458535 4730 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="4bd136f8-bcf2-4ddf-b885-d9455bd1b1a9" Feb 21 00:10:28 crc kubenswrapper[4730]: I0221 00:10:28.806466 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Feb 21 00:10:29 crc kubenswrapper[4730]: I0221 00:10:29.236542 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Feb 21 00:10:29 crc kubenswrapper[4730]: I0221 00:10:29.976367 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Feb 21 00:10:31 crc kubenswrapper[4730]: I0221 00:10:31.490750 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Feb 21 00:10:32 crc kubenswrapper[4730]: I0221 00:10:32.299002 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Feb 21 00:10:32 crc kubenswrapper[4730]: I0221 00:10:32.451053 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Feb 21 00:10:32 crc kubenswrapper[4730]: I0221 00:10:32.610882 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Feb 21 00:10:32 crc kubenswrapper[4730]: I0221 00:10:32.831996 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Feb 21 00:10:33 crc kubenswrapper[4730]: I0221 00:10:33.106892 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Feb 21 00:10:33 crc kubenswrapper[4730]: I0221 00:10:33.718187 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.023954 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.043834 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.120032 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.279631 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.539417 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.648546 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.710849 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.731446 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.772040 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.782202 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.811029 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.848681 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Feb 21 00:10:34 crc kubenswrapper[4730]: I0221 00:10:34.882714 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.178788 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.232340 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.492413 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.665862 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.865422 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Feb 21 00:10:35 crc kubenswrapper[4730]: I0221 00:10:35.970460 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.102083 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.163514 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.258065 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.308784 4730 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.329690 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.373933 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.434635 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.465715 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.539751 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.635052 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.788571 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.820191 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.836924 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.837347 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.853702 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.883596 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Feb 21 00:10:36 crc kubenswrapper[4730]: I0221 00:10:36.985705 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.022228 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.063477 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.088773 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.091004 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.091113 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.202378 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.251761 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.265419 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.266438 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.286032 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.309926 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.396444 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.583250 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.620076 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.659994 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.743534 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.768696 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.816983 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 21 00:10:37 crc kubenswrapper[4730]: I0221 00:10:37.997288 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.060220 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.063277 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.098741 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.213892 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.221205 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.225338 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.228387 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.252130 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.313710 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.435940 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.458110 4730 scope.go:117] "RemoveContainer" containerID="8b2a0ba92a6cea58b687f61b8ca716a2e782ad98635dcb1ed0cba58bcdfc87bd" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.524189 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.550274 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.671068 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.698737 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.716327 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.771976 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.812983 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.864666 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Feb 21 00:10:38 crc kubenswrapper[4730]: I0221 00:10:38.975665 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.004666 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.075146 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.145547 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.158130 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.158182 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06"} Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.236708 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.419643 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.566563 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.605265 4730 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.626513 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.638836 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.720876 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.947701 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.962839 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.963459 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Feb 21 00:10:39 crc kubenswrapper[4730]: I0221 00:10:39.986444 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.010836 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.075635 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.079726 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.099012 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.113889 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.156958 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.170194 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.170750 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/1.log" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.170830 4730 generic.go:334] "Generic (PLEG): container finished" podID="9d751cbb-f2e2-430d-9754-c882a5e924a5" containerID="69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06" exitCode=255 Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.170860 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerDied","Data":"69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06"} Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.170918 4730 scope.go:117] "RemoveContainer" containerID="8b2a0ba92a6cea58b687f61b8ca716a2e782ad98635dcb1ed0cba58bcdfc87bd" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.171658 4730 scope.go:117] "RemoveContainer" containerID="69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06" Feb 21 00:10:40 crc kubenswrapper[4730]: E0221 00:10:40.172104 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.291668 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.300705 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.457687 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.465701 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.475530 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.515311 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.611944 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.720568 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Feb 21 00:10:40 crc kubenswrapper[4730]: I0221 00:10:40.877089 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.048635 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.068464 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.077656 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.177690 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.259007 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.328098 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.512716 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.587305 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.601096 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.665302 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.735384 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.737610 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.770884 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.805235 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.842667 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.891507 4730 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.903160 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.905145 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Feb 21 00:10:41 crc kubenswrapper[4730]: I0221 00:10:41.923996 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.089652 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.093415 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.119018 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.155719 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.290474 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.309554 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.328235 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.436498 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.445266 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.575111 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.605108 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.622912 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.676596 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.791812 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.887515 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.903084 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.936535 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.971459 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Feb 21 00:10:42 crc kubenswrapper[4730]: I0221 00:10:42.971751 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.021029 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.037054 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.048873 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.065471 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.092657 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.153745 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.302127 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.316031 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.394276 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.496703 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.534991 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.619960 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.706621 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.754820 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Feb 21 00:10:43 crc kubenswrapper[4730]: I0221 00:10:43.946365 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.012322 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.062471 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.141062 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.189414 4730 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.190602 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podStartSLOduration=43.190575843 podStartE2EDuration="43.190575843s" podCreationTimestamp="2026-02-21 00:10:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:10:20.816551435 +0000 UTC m=+273.107387175" watchObservedRunningTime="2026-02-21 00:10:44.190575843 +0000 UTC m=+296.481411613" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.198403 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-ssd6c","openshift-kube-apiserver/kube-apiserver-crc"] Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.198507 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-b48d4c66-9sx76","openshift-kube-apiserver/kube-apiserver-crc"] Feb 21 00:10:44 crc kubenswrapper[4730]: E0221 00:10:44.199154 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" containerName="installer" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199191 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" containerName="installer" Feb 21 00:10:44 crc kubenswrapper[4730]: E0221 00:10:44.199215 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" containerName="oauth-openshift" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199223 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" containerName="oauth-openshift" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199363 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" containerName="oauth-openshift" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199383 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b903623-cae8-4c9e-ab8f-4fb507a3a320" containerName="installer" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199447 4730 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199486 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="aa2dfeb0-f137-4b5b-be18-5ea391e6b4e4" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.199884 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.204427 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.204981 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.205175 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.205538 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.205765 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.205943 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.206043 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.206244 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.206450 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.210809 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.210856 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.210974 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.211138 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.218410 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.220655 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.231170 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.231651 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=23.231622255 podStartE2EDuration="23.231622255s" podCreationTimestamp="2026-02-21 00:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:10:44.226531838 +0000 UTC m=+296.517367618" watchObservedRunningTime="2026-02-21 00:10:44.231622255 +0000 UTC m=+296.522458025" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.248546 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.347496 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.370932 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394705 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-service-ca\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394770 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394801 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394823 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g94h4\" (UniqueName: \"kubernetes.io/projected/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-kube-api-access-g94h4\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394853 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-login\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394877 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-dir\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394923 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394953 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-session\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.394986 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-policies\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.395006 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.395028 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.395063 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-router-certs\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.395087 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.395466 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-error\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.445213 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cf9c54fa-0743-40ed-85b7-ae38607f7265" path="/var/lib/kubelet/pods/cf9c54fa-0743-40ed-85b7-ae38607f7265/volumes" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.476411 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.478766 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.496862 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-error\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.496948 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-service-ca\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.496983 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497015 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497449 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g94h4\" (UniqueName: \"kubernetes.io/projected/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-kube-api-access-g94h4\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497525 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-login\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497565 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-dir\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497640 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497690 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-session\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497742 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-policies\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497771 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497827 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497873 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-router-certs\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.497936 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.498272 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-dir\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.499447 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-audit-policies\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.499468 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-service-ca\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.500309 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.500647 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-cliconfig\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.503085 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.504455 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-error\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.505778 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.506689 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.509356 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-user-template-login\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.509524 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.509660 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-router-certs\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.511338 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-serving-cert\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.512880 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-v4-0-config-system-session\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.522891 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g94h4\" (UniqueName: \"kubernetes.io/projected/6cd6bc77-d11d-4384-9bdd-3e37813cdcac-kube-api-access-g94h4\") pod \"oauth-openshift-b48d4c66-9sx76\" (UID: \"6cd6bc77-d11d-4384-9bdd-3e37813cdcac\") " pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.526765 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.538845 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.648314 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.657380 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.726489 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.841461 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.895782 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.906267 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Feb 21 00:10:44 crc kubenswrapper[4730]: I0221 00:10:44.929629 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.055734 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.084541 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.148101 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.184543 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.242682 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b48d4c66-9sx76"] Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.447469 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.581499 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.582137 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.647416 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.796872 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-b48d4c66-9sx76"] Feb 21 00:10:45 crc kubenswrapper[4730]: W0221 00:10:45.806889 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6cd6bc77_d11d_4384_9bdd_3e37813cdcac.slice/crio-e04926951e670ee869aa0a8cab7b50389b2f0f1ed2133a09979b03f1e8d4af12 WatchSource:0}: Error finding container e04926951e670ee869aa0a8cab7b50389b2f0f1ed2133a09979b03f1e8d4af12: Status 404 returned error can't find the container with id e04926951e670ee869aa0a8cab7b50389b2f0f1ed2133a09979b03f1e8d4af12 Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.896685 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.914572 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Feb 21 00:10:45 crc kubenswrapper[4730]: I0221 00:10:45.947731 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.121165 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.214464 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" event={"ID":"6cd6bc77-d11d-4384-9bdd-3e37813cdcac","Type":"ContainerStarted","Data":"95eec2427a7b87f40dcba6b033a2a6698c6e4ff3fe69cc0d87f761922b66827e"} Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.214516 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" event={"ID":"6cd6bc77-d11d-4384-9bdd-3e37813cdcac","Type":"ContainerStarted","Data":"e04926951e670ee869aa0a8cab7b50389b2f0f1ed2133a09979b03f1e8d4af12"} Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.215922 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.217662 4730 patch_prober.go:28] interesting pod/oauth-openshift-b48d4c66-9sx76 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": dial tcp 10.217.0.66:6443: connect: connection refused" start-of-body= Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.217721 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" podUID="6cd6bc77-d11d-4384-9bdd-3e37813cdcac" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": dial tcp 10.217.0.66:6443: connect: connection refused" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.220654 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.237113 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" podStartSLOduration=58.237092826 podStartE2EDuration="58.237092826s" podCreationTimestamp="2026-02-21 00:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:10:46.233257985 +0000 UTC m=+298.524093735" watchObservedRunningTime="2026-02-21 00:10:46.237092826 +0000 UTC m=+298.527928556" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.282999 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.315277 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.377221 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.488540 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.508504 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.633535 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Feb 21 00:10:46 crc kubenswrapper[4730]: I0221 00:10:46.791260 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.185679 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.211869 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.228554 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-b48d4c66-9sx76" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.253377 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.343698 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.373320 4730 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.390164 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.448102 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.484412 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.494732 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.626527 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.726005 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.758037 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.800824 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.822406 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Feb 21 00:10:47 crc kubenswrapper[4730]: I0221 00:10:47.910385 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.029271 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.069295 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.121065 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.205848 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.230851 4730 cert_rotation.go:91] certificate rotation detected, shutting down client connections to start using new credentials Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.336977 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.503009 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.545441 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.554266 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.605846 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.648550 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.750864 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.879340 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Feb 21 00:10:48 crc kubenswrapper[4730]: I0221 00:10:48.971349 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Feb 21 00:10:49 crc kubenswrapper[4730]: I0221 00:10:49.328684 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Feb 21 00:10:49 crc kubenswrapper[4730]: I0221 00:10:49.407665 4730 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Feb 21 00:10:50 crc kubenswrapper[4730]: I0221 00:10:50.144707 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 21 00:10:52 crc kubenswrapper[4730]: I0221 00:10:52.435177 4730 scope.go:117] "RemoveContainer" containerID="69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06" Feb 21 00:10:52 crc kubenswrapper[4730]: E0221 00:10:52.436595 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=check-endpoints pod=network-check-source-55646444c4-trplf_openshift-network-diagnostics(9d751cbb-f2e2-430d-9754-c882a5e924a5)\"" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.800802 4730 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.802218 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://37c6eb74e229c8794c48181f96eb406807a6a6bccbdd1e28a91cbe38147a06eb" gracePeriod=5 Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.967017 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.967363 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rk2d5" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="registry-server" containerID="cri-o://411493cf2a6ac06ae122658ad3f190a319226de5fa653da34272035f992d675d" gracePeriod=30 Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.973120 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.973504 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-d4xfg" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="registry-server" containerID="cri-o://c044e9003fbcd11ab085b351a35b65c6bd2f9566ef07b619d3fe0cde32f02229" gracePeriod=30 Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.986527 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.986709 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" podUID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" containerName="marketplace-operator" containerID="cri-o://a65d9e17aac5d697a17704052c8a3d83a53e3fddc489a0732b6323966fcff9ac" gracePeriod=30 Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.994690 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:10:54 crc kubenswrapper[4730]: I0221 00:10:54.994965 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5hhj6" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="registry-server" containerID="cri-o://a9d7dd6348892cd6c3793c2ee126f7b668e62b21b734935b136b35c8549a9cab" gracePeriod=30 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.003457 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rs768"] Feb 21 00:10:55 crc kubenswrapper[4730]: E0221 00:10:55.003726 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.003740 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.003888 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.004417 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.037257 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.037658 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4j2h7" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="registry-server" containerID="cri-o://b1924224a57f8dd76b64e50698b0a6fdbb722fc96e57b5460ff8e68e3c7dd8a1" gracePeriod=30 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.041042 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rs768"] Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.150337 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.150829 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ctxp\" (UniqueName: \"kubernetes.io/projected/6d866940-c241-40d7-b8e6-0a10dedbf9d1-kube-api-access-9ctxp\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.150876 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.251707 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ctxp\" (UniqueName: \"kubernetes.io/projected/6d866940-c241-40d7-b8e6-0a10dedbf9d1-kube-api-access-9ctxp\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.251753 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.251811 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.253892 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.257826 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6d866940-c241-40d7-b8e6-0a10dedbf9d1-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.271787 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ctxp\" (UniqueName: \"kubernetes.io/projected/6d866940-c241-40d7-b8e6-0a10dedbf9d1-kube-api-access-9ctxp\") pod \"marketplace-operator-79b997595-rs768\" (UID: \"6d866940-c241-40d7-b8e6-0a10dedbf9d1\") " pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.274671 4730 generic.go:334] "Generic (PLEG): container finished" podID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerID="411493cf2a6ac06ae122658ad3f190a319226de5fa653da34272035f992d675d" exitCode=0 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.274750 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerDied","Data":"411493cf2a6ac06ae122658ad3f190a319226de5fa653da34272035f992d675d"} Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.276610 4730 generic.go:334] "Generic (PLEG): container finished" podID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" containerID="a65d9e17aac5d697a17704052c8a3d83a53e3fddc489a0732b6323966fcff9ac" exitCode=0 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.276673 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" event={"ID":"6c73c5be-27dd-4e76-9ad9-9008fd2a884d","Type":"ContainerDied","Data":"a65d9e17aac5d697a17704052c8a3d83a53e3fddc489a0732b6323966fcff9ac"} Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.279828 4730 generic.go:334] "Generic (PLEG): container finished" podID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerID="b1924224a57f8dd76b64e50698b0a6fdbb722fc96e57b5460ff8e68e3c7dd8a1" exitCode=0 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.279932 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerDied","Data":"b1924224a57f8dd76b64e50698b0a6fdbb722fc96e57b5460ff8e68e3c7dd8a1"} Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.282159 4730 generic.go:334] "Generic (PLEG): container finished" podID="17855419-b9b5-4c88-a004-694eb8320ea7" containerID="c044e9003fbcd11ab085b351a35b65c6bd2f9566ef07b619d3fe0cde32f02229" exitCode=0 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.282221 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerDied","Data":"c044e9003fbcd11ab085b351a35b65c6bd2f9566ef07b619d3fe0cde32f02229"} Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.284069 4730 generic.go:334] "Generic (PLEG): container finished" podID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerID="a9d7dd6348892cd6c3793c2ee126f7b668e62b21b734935b136b35c8549a9cab" exitCode=0 Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.284110 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerDied","Data":"a9d7dd6348892cd6c3793c2ee126f7b668e62b21b734935b136b35c8549a9cab"} Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.321665 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.473037 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.495481 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.516860 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.534922 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.580985 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.655819 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8vtzm\" (UniqueName: \"kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm\") pod \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.655927 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities\") pod \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.656006 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drr9w\" (UniqueName: \"kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w\") pod \"17855419-b9b5-4c88-a004-694eb8320ea7\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.656100 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s62rx\" (UniqueName: \"kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx\") pod \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.656149 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content\") pod \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\" (UID: \"d8b1b254-42f2-4ff9-8971-13528ab4f8d0\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.656255 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities\") pod \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.656770 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities" (OuterVolumeSpecName: "utilities") pod "d8b1b254-42f2-4ff9-8971-13528ab4f8d0" (UID: "d8b1b254-42f2-4ff9-8971-13528ab4f8d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670664 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics\") pod \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670726 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tjxkl\" (UniqueName: \"kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl\") pod \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670762 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca\") pod \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\" (UID: \"6c73c5be-27dd-4e76-9ad9-9008fd2a884d\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670793 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj4zb\" (UniqueName: \"kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb\") pod \"789966a0-6952-4ab6-9baf-cd53e5c06270\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670840 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities\") pod \"789966a0-6952-4ab6-9baf-cd53e5c06270\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670872 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content\") pod \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\" (UID: \"063a7dbe-6af4-4aff-9d3a-6f4394b98366\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670927 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities\") pod \"17855419-b9b5-4c88-a004-694eb8320ea7\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.670983 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content\") pod \"17855419-b9b5-4c88-a004-694eb8320ea7\" (UID: \"17855419-b9b5-4c88-a004-694eb8320ea7\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.671028 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content\") pod \"789966a0-6952-4ab6-9baf-cd53e5c06270\" (UID: \"789966a0-6952-4ab6-9baf-cd53e5c06270\") " Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.671399 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.657497 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities" (OuterVolumeSpecName: "utilities") pod "063a7dbe-6af4-4aff-9d3a-6f4394b98366" (UID: "063a7dbe-6af4-4aff-9d3a-6f4394b98366"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.659749 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx" (OuterVolumeSpecName: "kube-api-access-s62rx") pod "063a7dbe-6af4-4aff-9d3a-6f4394b98366" (UID: "063a7dbe-6af4-4aff-9d3a-6f4394b98366"). InnerVolumeSpecName "kube-api-access-s62rx". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.659985 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w" (OuterVolumeSpecName: "kube-api-access-drr9w") pod "17855419-b9b5-4c88-a004-694eb8320ea7" (UID: "17855419-b9b5-4c88-a004-694eb8320ea7"). InnerVolumeSpecName "kube-api-access-drr9w". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.662333 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm" (OuterVolumeSpecName: "kube-api-access-8vtzm") pod "d8b1b254-42f2-4ff9-8971-13528ab4f8d0" (UID: "d8b1b254-42f2-4ff9-8971-13528ab4f8d0"). InnerVolumeSpecName "kube-api-access-8vtzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.671441 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "6c73c5be-27dd-4e76-9ad9-9008fd2a884d" (UID: "6c73c5be-27dd-4e76-9ad9-9008fd2a884d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.671542 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities" (OuterVolumeSpecName: "utilities") pod "789966a0-6952-4ab6-9baf-cd53e5c06270" (UID: "789966a0-6952-4ab6-9baf-cd53e5c06270"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.672368 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities" (OuterVolumeSpecName: "utilities") pod "17855419-b9b5-4c88-a004-694eb8320ea7" (UID: "17855419-b9b5-4c88-a004-694eb8320ea7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.674157 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "6c73c5be-27dd-4e76-9ad9-9008fd2a884d" (UID: "6c73c5be-27dd-4e76-9ad9-9008fd2a884d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.676635 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl" (OuterVolumeSpecName: "kube-api-access-tjxkl") pod "6c73c5be-27dd-4e76-9ad9-9008fd2a884d" (UID: "6c73c5be-27dd-4e76-9ad9-9008fd2a884d"). InnerVolumeSpecName "kube-api-access-tjxkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.677132 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb" (OuterVolumeSpecName: "kube-api-access-cj4zb") pod "789966a0-6952-4ab6-9baf-cd53e5c06270" (UID: "789966a0-6952-4ab6-9baf-cd53e5c06270"). InnerVolumeSpecName "kube-api-access-cj4zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.699732 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "063a7dbe-6af4-4aff-9d3a-6f4394b98366" (UID: "063a7dbe-6af4-4aff-9d3a-6f4394b98366"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.727206 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d8b1b254-42f2-4ff9-8971-13528ab4f8d0" (UID: "d8b1b254-42f2-4ff9-8971-13528ab4f8d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.742569 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "17855419-b9b5-4c88-a004-694eb8320ea7" (UID: "17855419-b9b5-4c88-a004-694eb8320ea7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772599 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772633 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8vtzm\" (UniqueName: \"kubernetes.io/projected/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-kube-api-access-8vtzm\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772644 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drr9w\" (UniqueName: \"kubernetes.io/projected/17855419-b9b5-4c88-a004-694eb8320ea7-kube-api-access-drr9w\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772652 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s62rx\" (UniqueName: \"kubernetes.io/projected/063a7dbe-6af4-4aff-9d3a-6f4394b98366-kube-api-access-s62rx\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772662 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d8b1b254-42f2-4ff9-8971-13528ab4f8d0-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772671 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772679 4730 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772689 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tjxkl\" (UniqueName: \"kubernetes.io/projected/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-kube-api-access-tjxkl\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772701 4730 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6c73c5be-27dd-4e76-9ad9-9008fd2a884d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772709 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj4zb\" (UniqueName: \"kubernetes.io/projected/789966a0-6952-4ab6-9baf-cd53e5c06270-kube-api-access-cj4zb\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772717 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772725 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/063a7dbe-6af4-4aff-9d3a-6f4394b98366-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.772735 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/17855419-b9b5-4c88-a004-694eb8320ea7-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.824820 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "789966a0-6952-4ab6-9baf-cd53e5c06270" (UID: "789966a0-6952-4ab6-9baf-cd53e5c06270"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.833099 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rs768"] Feb 21 00:10:55 crc kubenswrapper[4730]: I0221 00:10:55.874113 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/789966a0-6952-4ab6-9baf-cd53e5c06270-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.291977 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.293047 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-vbnbc" event={"ID":"6c73c5be-27dd-4e76-9ad9-9008fd2a884d","Type":"ContainerDied","Data":"0dcb8984b8f1c2fa1671feabafb47b411cab0f37daaab4ef4c1a8093b9b58a53"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.293172 4730 scope.go:117] "RemoveContainer" containerID="a65d9e17aac5d697a17704052c8a3d83a53e3fddc489a0732b6323966fcff9ac" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.296681 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4j2h7" event={"ID":"789966a0-6952-4ab6-9baf-cd53e5c06270","Type":"ContainerDied","Data":"fdf65f079d4e0207471013613fdb0235d927ef1c9c284a5ed82ed544712e9f8b"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.296819 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4j2h7" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.305342 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-d4xfg" event={"ID":"17855419-b9b5-4c88-a004-694eb8320ea7","Type":"ContainerDied","Data":"3972471d1a40ed3de9c1598bafd20ed04872fc57aa083acb76e6889c2d562eb5"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.305505 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-d4xfg" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.311477 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5hhj6" event={"ID":"063a7dbe-6af4-4aff-9d3a-6f4394b98366","Type":"ContainerDied","Data":"222f3f7bd6c6e2f97c2bfa1d995846f8032a2fa1c8664c2bbc2409cf5b245a03"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.311612 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5hhj6" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.318697 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" event={"ID":"6d866940-c241-40d7-b8e6-0a10dedbf9d1","Type":"ContainerStarted","Data":"ada50d331da7eaa4bdd4254b0342d7503983426efd7c6d0f286c9ac4017ef5e5"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.318766 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" event={"ID":"6d866940-c241-40d7-b8e6-0a10dedbf9d1","Type":"ContainerStarted","Data":"649b60194232b387bcaf3dc71ed9932943ad34e53b963bf8913a19f03951acf4"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.320356 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.320504 4730 scope.go:117] "RemoveContainer" containerID="b1924224a57f8dd76b64e50698b0a6fdbb722fc96e57b5460ff8e68e3c7dd8a1" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.322077 4730 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rs768 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.67:8080/healthz\": dial tcp 10.217.0.67:8080: connect: connection refused" start-of-body= Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.322138 4730 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" podUID="6d866940-c241-40d7-b8e6-0a10dedbf9d1" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.67:8080/healthz\": dial tcp 10.217.0.67:8080: connect: connection refused" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.328944 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rk2d5" event={"ID":"d8b1b254-42f2-4ff9-8971-13528ab4f8d0","Type":"ContainerDied","Data":"79f0b515be9b1050e7f9e9c36c2d89d8a44277033f208a47aa8716e3a19a828d"} Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.329104 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rk2d5" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.344412 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.355186 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-vbnbc"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.361111 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" podStartSLOduration=2.361095718 podStartE2EDuration="2.361095718s" podCreationTimestamp="2026-02-21 00:10:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:10:56.356510305 +0000 UTC m=+308.647346045" watchObservedRunningTime="2026-02-21 00:10:56.361095718 +0000 UTC m=+308.651931448" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.377355 4730 scope.go:117] "RemoveContainer" containerID="623f80859bfcb4356487021bf1a827e996e7ddd6a14c82ab46345c76f31c0805" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.386016 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.392193 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rk2d5"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.396039 4730 scope.go:117] "RemoveContainer" containerID="20ec430915a54169ae28cee1bd4eb61951bc732541b2f3d339c4c6b8ec7e3861" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.398087 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.404954 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-d4xfg"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.410880 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.415545 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4j2h7"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.420891 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.420957 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5hhj6"] Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.427320 4730 scope.go:117] "RemoveContainer" containerID="c044e9003fbcd11ab085b351a35b65c6bd2f9566ef07b619d3fe0cde32f02229" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.442825 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" path="/var/lib/kubelet/pods/063a7dbe-6af4-4aff-9d3a-6f4394b98366/volumes" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.444004 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" path="/var/lib/kubelet/pods/17855419-b9b5-4c88-a004-694eb8320ea7/volumes" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.445044 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" path="/var/lib/kubelet/pods/6c73c5be-27dd-4e76-9ad9-9008fd2a884d/volumes" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.447133 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" path="/var/lib/kubelet/pods/789966a0-6952-4ab6-9baf-cd53e5c06270/volumes" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.449797 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" path="/var/lib/kubelet/pods/d8b1b254-42f2-4ff9-8971-13528ab4f8d0/volumes" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.451946 4730 scope.go:117] "RemoveContainer" containerID="d77998f1f3e84c55d274bf7d4ab13c2ac4787cac5e7d55ba4a92f8d1516cdf27" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.468173 4730 scope.go:117] "RemoveContainer" containerID="10110d5987eddb8dfbf41b31795b16aa36eff1fee09ea3651f5190afd20fcba6" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.488426 4730 scope.go:117] "RemoveContainer" containerID="a9d7dd6348892cd6c3793c2ee126f7b668e62b21b734935b136b35c8549a9cab" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.505299 4730 scope.go:117] "RemoveContainer" containerID="91de3cddfcbe8b1005c4de6537c6ca655f7008d41555ab3a3813e4f2268cee28" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.518316 4730 scope.go:117] "RemoveContainer" containerID="7c2b7d1f20613c9af58bd7dae231689b8e72c73cdeea040e7f0ed220babf5798" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.529930 4730 scope.go:117] "RemoveContainer" containerID="411493cf2a6ac06ae122658ad3f190a319226de5fa653da34272035f992d675d" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.548349 4730 scope.go:117] "RemoveContainer" containerID="ab6d5ba50a7611477525bd9fc7feabc5770b79f7d4545b84878a26cd61308f5c" Feb 21 00:10:56 crc kubenswrapper[4730]: I0221 00:10:56.563102 4730 scope.go:117] "RemoveContainer" containerID="9d4f5a2a9be586fcd8f0e65cf0835f92dc5b6f3692f16212dda3b235c49e2b57" Feb 21 00:10:57 crc kubenswrapper[4730]: I0221 00:10:57.351192 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rs768" Feb 21 00:10:58 crc kubenswrapper[4730]: I0221 00:10:58.790302 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.358293 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.358682 4730 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="37c6eb74e229c8794c48181f96eb406807a6a6bccbdd1e28a91cbe38147a06eb" exitCode=137 Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.358734 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c55f741074fc910be05bc1e509ebf33f1a5e4aa9bbe3ec1feed30d116fd716eb" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.376388 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.376477 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.443358 4730 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.453091 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.453126 4730 kubelet.go:2649] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="2a2741a6-e1aa-4e29-9577-e2f2364e49b1" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.466395 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.466453 4730 kubelet.go:2673] "Unable to find pod for mirror pod, skipping" mirrorPod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" mirrorPodUID="2a2741a6-e1aa-4e29-9577-e2f2364e49b1" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469380 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469519 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469559 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469638 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469686 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469520 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469814 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469825 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.469875 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.470064 4730 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.470086 4730 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.470095 4730 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.470104 4730 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.477396 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:11:00 crc kubenswrapper[4730]: I0221 00:11:00.571641 4730 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Feb 21 00:11:01 crc kubenswrapper[4730]: I0221 00:11:01.364013 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Feb 21 00:11:02 crc kubenswrapper[4730]: I0221 00:11:02.444754 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Feb 21 00:11:03 crc kubenswrapper[4730]: I0221 00:11:03.435671 4730 scope.go:117] "RemoveContainer" containerID="69b628fb22430c5c37d46681df418bce86cda2a35abe61ae91dd4b4f3c1b5b06" Feb 21 00:11:04 crc kubenswrapper[4730]: I0221 00:11:04.385064 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-network-diagnostics_network-check-source-55646444c4-trplf_9d751cbb-f2e2-430d-9754-c882a5e924a5/check-endpoints/2.log" Feb 21 00:11:04 crc kubenswrapper[4730]: I0221 00:11:04.385508 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"8bd45984d05b0ff8dc42cc1b8852e2ab8252d0d1d368e8e297c1886a3418a0ed"} Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.882631 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g5wkr"] Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883640 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883663 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883682 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883694 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883712 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883725 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883750 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883762 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883779 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883793 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883819 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883834 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883869 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.883885 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.883999 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.884016 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.884147 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.884162 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.886033 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886091 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.886116 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" containerName="marketplace-operator" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886130 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" containerName="marketplace-operator" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.886150 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886163 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="extract-utilities" Feb 21 00:11:46 crc kubenswrapper[4730]: E0221 00:11:46.886187 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886202 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="extract-content" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886478 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="17855419-b9b5-4c88-a004-694eb8320ea7" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886499 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8b1b254-42f2-4ff9-8971-13528ab4f8d0" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886516 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c73c5be-27dd-4e76-9ad9-9008fd2a884d" containerName="marketplace-operator" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886537 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="063a7dbe-6af4-4aff-9d3a-6f4394b98366" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.886551 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="789966a0-6952-4ab6-9baf-cd53e5c06270" containerName="registry-server" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.887986 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.899091 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.902615 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5wkr"] Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.982063 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-catalog-content\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.982166 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqklb\" (UniqueName: \"kubernetes.io/projected/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-kube-api-access-kqklb\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:46 crc kubenswrapper[4730]: I0221 00:11:46.982409 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-utilities\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.074436 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-tr8gl"] Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.076111 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.079440 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084360 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-utilities\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084443 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-catalog-content\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084560 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-utilities\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084637 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wmfr\" (UniqueName: \"kubernetes.io/projected/308c905e-1c42-49c8-8758-67231973e7a8-kube-api-access-6wmfr\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084684 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-catalog-content\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.084723 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqklb\" (UniqueName: \"kubernetes.io/projected/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-kube-api-access-kqklb\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.085425 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-utilities\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.085537 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-catalog-content\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.100094 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr8gl"] Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.121295 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqklb\" (UniqueName: \"kubernetes.io/projected/f981a395-9367-4b4b-9ef8-d3aeee5a65f3-kube-api-access-kqklb\") pod \"redhat-operators-g5wkr\" (UID: \"f981a395-9367-4b4b-9ef8-d3aeee5a65f3\") " pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.185444 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wmfr\" (UniqueName: \"kubernetes.io/projected/308c905e-1c42-49c8-8758-67231973e7a8-kube-api-access-6wmfr\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.185539 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-utilities\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.185577 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-catalog-content\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.186182 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-catalog-content\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.186291 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/308c905e-1c42-49c8-8758-67231973e7a8-utilities\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.207196 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wmfr\" (UniqueName: \"kubernetes.io/projected/308c905e-1c42-49c8-8758-67231973e7a8-kube-api-access-6wmfr\") pod \"community-operators-tr8gl\" (UID: \"308c905e-1c42-49c8-8758-67231973e7a8\") " pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.220606 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.471534 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.611431 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g5wkr"] Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.653332 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5wkr" event={"ID":"f981a395-9367-4b4b-9ef8-d3aeee5a65f3","Type":"ContainerStarted","Data":"ae78cd9c6d468eb2d8f99112707c6a4e02e385dc15b1f527e56c4a6b811eb763"} Feb 21 00:11:47 crc kubenswrapper[4730]: I0221 00:11:47.905277 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-tr8gl"] Feb 21 00:11:47 crc kubenswrapper[4730]: W0221 00:11:47.914114 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod308c905e_1c42_49c8_8758_67231973e7a8.slice/crio-ecb656c6639c6d4127d3ac0cce33396344e148652fa1b6af58488883206f6afe WatchSource:0}: Error finding container ecb656c6639c6d4127d3ac0cce33396344e148652fa1b6af58488883206f6afe: Status 404 returned error can't find the container with id ecb656c6639c6d4127d3ac0cce33396344e148652fa1b6af58488883206f6afe Feb 21 00:11:48 crc kubenswrapper[4730]: I0221 00:11:48.663558 4730 generic.go:334] "Generic (PLEG): container finished" podID="f981a395-9367-4b4b-9ef8-d3aeee5a65f3" containerID="4dd85fd87236f05d3c5afaeab87796da129cc41362140f082fe8741a10759660" exitCode=0 Feb 21 00:11:48 crc kubenswrapper[4730]: I0221 00:11:48.663708 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5wkr" event={"ID":"f981a395-9367-4b4b-9ef8-d3aeee5a65f3","Type":"ContainerDied","Data":"4dd85fd87236f05d3c5afaeab87796da129cc41362140f082fe8741a10759660"} Feb 21 00:11:48 crc kubenswrapper[4730]: I0221 00:11:48.667630 4730 generic.go:334] "Generic (PLEG): container finished" podID="308c905e-1c42-49c8-8758-67231973e7a8" containerID="5c6e4c937d784846e41e9108952266a45cb6ddc2b56c6560faf15c9905105bc2" exitCode=0 Feb 21 00:11:48 crc kubenswrapper[4730]: I0221 00:11:48.667665 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr8gl" event={"ID":"308c905e-1c42-49c8-8758-67231973e7a8","Type":"ContainerDied","Data":"5c6e4c937d784846e41e9108952266a45cb6ddc2b56c6560faf15c9905105bc2"} Feb 21 00:11:48 crc kubenswrapper[4730]: I0221 00:11:48.667681 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr8gl" event={"ID":"308c905e-1c42-49c8-8758-67231973e7a8","Type":"ContainerStarted","Data":"ecb656c6639c6d4127d3ac0cce33396344e148652fa1b6af58488883206f6afe"} Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.467426 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lwqnm"] Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.470224 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.474644 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.477669 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwqnm"] Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.528770 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-utilities\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.528833 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-catalog-content\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.528871 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2mghf\" (UniqueName: \"kubernetes.io/projected/f5ea3922-1e00-45de-9b77-881b5f8efe58-kube-api-access-2mghf\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.630017 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-utilities\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.630070 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-catalog-content\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.630129 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2mghf\" (UniqueName: \"kubernetes.io/projected/f5ea3922-1e00-45de-9b77-881b5f8efe58-kube-api-access-2mghf\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.630953 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-utilities\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.631310 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5ea3922-1e00-45de-9b77-881b5f8efe58-catalog-content\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.667654 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2mghf\" (UniqueName: \"kubernetes.io/projected/f5ea3922-1e00-45de-9b77-881b5f8efe58-kube-api-access-2mghf\") pod \"certified-operators-lwqnm\" (UID: \"f5ea3922-1e00-45de-9b77-881b5f8efe58\") " pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.681662 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.682804 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.683493 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5wkr" event={"ID":"f981a395-9367-4b4b-9ef8-d3aeee5a65f3","Type":"ContainerStarted","Data":"7837fdfd46f2284fd013d9f520f7860d30fa1869a46a42f238df3b7904c7fca1"} Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.685656 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.695221 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.730804 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.730859 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mzzq\" (UniqueName: \"kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.730936 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.791440 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.831665 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.831829 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.831874 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mzzq\" (UniqueName: \"kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.832594 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.833313 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:49 crc kubenswrapper[4730]: I0221 00:11:49.866452 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mzzq\" (UniqueName: \"kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq\") pod \"redhat-marketplace-8ghsj\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.044927 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwqnm"] Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.060225 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:11:50 crc kubenswrapper[4730]: W0221 00:11:50.060710 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf5ea3922_1e00_45de_9b77_881b5f8efe58.slice/crio-3928de0a76ab596b3ec379b8932c05e935b250febe53770d41d3e77c0dea2f75 WatchSource:0}: Error finding container 3928de0a76ab596b3ec379b8932c05e935b250febe53770d41d3e77c0dea2f75: Status 404 returned error can't find the container with id 3928de0a76ab596b3ec379b8932c05e935b250febe53770d41d3e77c0dea2f75 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.267374 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:11:50 crc kubenswrapper[4730]: W0221 00:11:50.288479 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55765fd7_a04e_4509_9fc6_2633eb09341b.slice/crio-1bf28d76776b06a993d4eaac574311f141f2b54c548eeb7173f07a7b61c94360 WatchSource:0}: Error finding container 1bf28d76776b06a993d4eaac574311f141f2b54c548eeb7173f07a7b61c94360: Status 404 returned error can't find the container with id 1bf28d76776b06a993d4eaac574311f141f2b54c548eeb7173f07a7b61c94360 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.692744 4730 generic.go:334] "Generic (PLEG): container finished" podID="308c905e-1c42-49c8-8758-67231973e7a8" containerID="e0cdf251ada29412b5d0f365b286ba574e6ca4148b60427537c0402d2c2f1f2a" exitCode=0 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.692936 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr8gl" event={"ID":"308c905e-1c42-49c8-8758-67231973e7a8","Type":"ContainerDied","Data":"e0cdf251ada29412b5d0f365b286ba574e6ca4148b60427537c0402d2c2f1f2a"} Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.696387 4730 generic.go:334] "Generic (PLEG): container finished" podID="f5ea3922-1e00-45de-9b77-881b5f8efe58" containerID="83cb5c1e4875f2684de54bb0f8bf536563bca977e17ede55029c6af04475702b" exitCode=0 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.696477 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwqnm" event={"ID":"f5ea3922-1e00-45de-9b77-881b5f8efe58","Type":"ContainerDied","Data":"83cb5c1e4875f2684de54bb0f8bf536563bca977e17ede55029c6af04475702b"} Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.696515 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwqnm" event={"ID":"f5ea3922-1e00-45de-9b77-881b5f8efe58","Type":"ContainerStarted","Data":"3928de0a76ab596b3ec379b8932c05e935b250febe53770d41d3e77c0dea2f75"} Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.698830 4730 generic.go:334] "Generic (PLEG): container finished" podID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerID="91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae" exitCode=0 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.698871 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerDied","Data":"91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae"} Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.698926 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerStarted","Data":"1bf28d76776b06a993d4eaac574311f141f2b54c548eeb7173f07a7b61c94360"} Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.704525 4730 generic.go:334] "Generic (PLEG): container finished" podID="f981a395-9367-4b4b-9ef8-d3aeee5a65f3" containerID="7837fdfd46f2284fd013d9f520f7860d30fa1869a46a42f238df3b7904c7fca1" exitCode=0 Feb 21 00:11:50 crc kubenswrapper[4730]: I0221 00:11:50.704554 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5wkr" event={"ID":"f981a395-9367-4b4b-9ef8-d3aeee5a65f3","Type":"ContainerDied","Data":"7837fdfd46f2284fd013d9f520f7860d30fa1869a46a42f238df3b7904c7fca1"} Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.714136 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g5wkr" event={"ID":"f981a395-9367-4b4b-9ef8-d3aeee5a65f3","Type":"ContainerStarted","Data":"4dd075855c3012b013cbb220b491f477391e29bbffca3559ce1f2dfd9984f792"} Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.716396 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-tr8gl" event={"ID":"308c905e-1c42-49c8-8758-67231973e7a8","Type":"ContainerStarted","Data":"892197a06c452409252bd760af62c5d4453ae3da407a6253df00b91b25aa30aa"} Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.722680 4730 generic.go:334] "Generic (PLEG): container finished" podID="f5ea3922-1e00-45de-9b77-881b5f8efe58" containerID="b8dbdf900dba4d26bc260915476791b67c01e83e4d7109786edd0e1419b85295" exitCode=0 Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.722737 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwqnm" event={"ID":"f5ea3922-1e00-45de-9b77-881b5f8efe58","Type":"ContainerDied","Data":"b8dbdf900dba4d26bc260915476791b67c01e83e4d7109786edd0e1419b85295"} Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.725522 4730 generic.go:334] "Generic (PLEG): container finished" podID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerID="81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b" exitCode=0 Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.725564 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerDied","Data":"81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b"} Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.765005 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g5wkr" podStartSLOduration=3.343204835 podStartE2EDuration="5.764987719s" podCreationTimestamp="2026-02-21 00:11:46 +0000 UTC" firstStartedPulling="2026-02-21 00:11:48.673176509 +0000 UTC m=+360.964012239" lastFinishedPulling="2026-02-21 00:11:51.094959363 +0000 UTC m=+363.385795123" observedRunningTime="2026-02-21 00:11:51.740875179 +0000 UTC m=+364.031710909" watchObservedRunningTime="2026-02-21 00:11:51.764987719 +0000 UTC m=+364.055823459" Feb 21 00:11:51 crc kubenswrapper[4730]: I0221 00:11:51.765541 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-tr8gl" podStartSLOduration=2.304151852 podStartE2EDuration="4.765533165s" podCreationTimestamp="2026-02-21 00:11:47 +0000 UTC" firstStartedPulling="2026-02-21 00:11:48.672712746 +0000 UTC m=+360.963548476" lastFinishedPulling="2026-02-21 00:11:51.134094019 +0000 UTC m=+363.424929789" observedRunningTime="2026-02-21 00:11:51.762669652 +0000 UTC m=+364.053505412" watchObservedRunningTime="2026-02-21 00:11:51.765533165 +0000 UTC m=+364.056368895" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.475813 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qvdzm"] Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.476920 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.489628 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qvdzm"] Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676233 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-trusted-ca\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676333 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-bound-sa-token\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676421 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676456 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-certificates\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676486 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03582224-6fbb-4c0d-8f81-2524fa14940c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676531 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvr6w\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-kube-api-access-hvr6w\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676567 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03582224-6fbb-4c0d-8f81-2524fa14940c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.676780 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-tls\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.707830 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.736026 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwqnm" event={"ID":"f5ea3922-1e00-45de-9b77-881b5f8efe58","Type":"ContainerStarted","Data":"6bfe424619eff4fc3d5c09066825794e8ff702b1c5754427758d6cdaca9e9918"} Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.739551 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerStarted","Data":"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185"} Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.758340 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lwqnm" podStartSLOduration=2.363509701 podStartE2EDuration="3.758326294s" podCreationTimestamp="2026-02-21 00:11:49 +0000 UTC" firstStartedPulling="2026-02-21 00:11:50.698034317 +0000 UTC m=+362.988870047" lastFinishedPulling="2026-02-21 00:11:52.09285091 +0000 UTC m=+364.383686640" observedRunningTime="2026-02-21 00:11:52.755553123 +0000 UTC m=+365.046388853" watchObservedRunningTime="2026-02-21 00:11:52.758326294 +0000 UTC m=+365.049162024" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.775202 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8ghsj" podStartSLOduration=2.35521722 podStartE2EDuration="3.775187114s" podCreationTimestamp="2026-02-21 00:11:49 +0000 UTC" firstStartedPulling="2026-02-21 00:11:50.702145166 +0000 UTC m=+362.992980896" lastFinishedPulling="2026-02-21 00:11:52.12211506 +0000 UTC m=+364.412950790" observedRunningTime="2026-02-21 00:11:52.773575707 +0000 UTC m=+365.064411427" watchObservedRunningTime="2026-02-21 00:11:52.775187114 +0000 UTC m=+365.066022834" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778182 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-bound-sa-token\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778269 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-certificates\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778300 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03582224-6fbb-4c0d-8f81-2524fa14940c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778341 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvr6w\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-kube-api-access-hvr6w\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778370 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03582224-6fbb-4c0d-8f81-2524fa14940c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778396 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-tls\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.778427 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-trusted-ca\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.779037 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/03582224-6fbb-4c0d-8f81-2524fa14940c-ca-trust-extracted\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.779560 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-certificates\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.781077 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/03582224-6fbb-4c0d-8f81-2524fa14940c-trusted-ca\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.784649 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/03582224-6fbb-4c0d-8f81-2524fa14940c-installation-pull-secrets\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.796924 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvr6w\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-kube-api-access-hvr6w\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.800626 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-registry-tls\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:52 crc kubenswrapper[4730]: I0221 00:11:52.804817 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/03582224-6fbb-4c0d-8f81-2524fa14940c-bound-sa-token\") pod \"image-registry-66df7c8f76-qvdzm\" (UID: \"03582224-6fbb-4c0d-8f81-2524fa14940c\") " pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:53 crc kubenswrapper[4730]: I0221 00:11:53.092803 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:53 crc kubenswrapper[4730]: I0221 00:11:53.324428 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-qvdzm"] Feb 21 00:11:53 crc kubenswrapper[4730]: W0221 00:11:53.327459 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod03582224_6fbb_4c0d_8f81_2524fa14940c.slice/crio-25aa118fa8a93737313aa89ea6fedcabf5e34ce87eac16b06b4812b0b9852732 WatchSource:0}: Error finding container 25aa118fa8a93737313aa89ea6fedcabf5e34ce87eac16b06b4812b0b9852732: Status 404 returned error can't find the container with id 25aa118fa8a93737313aa89ea6fedcabf5e34ce87eac16b06b4812b0b9852732 Feb 21 00:11:53 crc kubenswrapper[4730]: I0221 00:11:53.747238 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" event={"ID":"03582224-6fbb-4c0d-8f81-2524fa14940c","Type":"ContainerStarted","Data":"29982093c278cff92c5d943919ccc576a3daceac9bdd462e8c430ab9078ef38f"} Feb 21 00:11:53 crc kubenswrapper[4730]: I0221 00:11:53.747271 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" event={"ID":"03582224-6fbb-4c0d-8f81-2524fa14940c","Type":"ContainerStarted","Data":"25aa118fa8a93737313aa89ea6fedcabf5e34ce87eac16b06b4812b0b9852732"} Feb 21 00:11:53 crc kubenswrapper[4730]: I0221 00:11:53.747606 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.221229 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.221291 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.472716 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.472770 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.515565 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.533262 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" podStartSLOduration=5.533243117 podStartE2EDuration="5.533243117s" podCreationTimestamp="2026-02-21 00:11:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:11:53.763865612 +0000 UTC m=+366.054701342" watchObservedRunningTime="2026-02-21 00:11:57.533243117 +0000 UTC m=+369.824078867" Feb 21 00:11:57 crc kubenswrapper[4730]: I0221 00:11:57.806369 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-tr8gl" Feb 21 00:11:58 crc kubenswrapper[4730]: I0221 00:11:58.273757 4730 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-g5wkr" podUID="f981a395-9367-4b4b-9ef8-d3aeee5a65f3" containerName="registry-server" probeResult="failure" output=< Feb 21 00:11:58 crc kubenswrapper[4730]: timeout: failed to connect service ":50051" within 1s Feb 21 00:11:58 crc kubenswrapper[4730]: > Feb 21 00:11:59 crc kubenswrapper[4730]: I0221 00:11:59.792375 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:59 crc kubenswrapper[4730]: I0221 00:11:59.792426 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:11:59 crc kubenswrapper[4730]: I0221 00:11:59.847702 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:12:00 crc kubenswrapper[4730]: I0221 00:12:00.061467 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:12:00 crc kubenswrapper[4730]: I0221 00:12:00.061861 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:12:00 crc kubenswrapper[4730]: I0221 00:12:00.109447 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:12:00 crc kubenswrapper[4730]: I0221 00:12:00.867703 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:12:00 crc kubenswrapper[4730]: I0221 00:12:00.869535 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lwqnm" Feb 21 00:12:05 crc kubenswrapper[4730]: I0221 00:12:05.775477 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:12:05 crc kubenswrapper[4730]: I0221 00:12:05.776199 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:12:07 crc kubenswrapper[4730]: I0221 00:12:07.267250 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:12:07 crc kubenswrapper[4730]: I0221 00:12:07.305444 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g5wkr" Feb 21 00:12:13 crc kubenswrapper[4730]: I0221 00:12:13.099414 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-qvdzm" Feb 21 00:12:13 crc kubenswrapper[4730]: I0221 00:12:13.151815 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:12:35 crc kubenswrapper[4730]: I0221 00:12:35.776082 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:12:35 crc kubenswrapper[4730]: I0221 00:12:35.777952 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.221619 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" podUID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" containerName="registry" containerID="cri-o://fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd" gracePeriod=30 Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.681140 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.804221 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.804543 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld7h5\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.804639 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.804850 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.804887 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.806029 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.806054 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.806106 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.806185 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.806253 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted\") pod \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\" (UID: \"691c1af1-e7b5-4b53-8f4d-9a33e48106be\") " Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.815440 4730 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-certificates\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.815481 4730 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/691c1af1-e7b5-4b53-8f4d-9a33e48106be-trusted-ca\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.819101 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.819160 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.819421 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5" (OuterVolumeSpecName: "kube-api-access-ld7h5") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "kube-api-access-ld7h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.819948 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.821364 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.830879 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "691c1af1-e7b5-4b53-8f4d-9a33e48106be" (UID: "691c1af1-e7b5-4b53-8f4d-9a33e48106be"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.916816 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld7h5\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-kube-api-access-ld7h5\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.916855 4730 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-bound-sa-token\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.916867 4730 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/691c1af1-e7b5-4b53-8f4d-9a33e48106be-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.916878 4730 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/691c1af1-e7b5-4b53-8f4d-9a33e48106be-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:38 crc kubenswrapper[4730]: I0221 00:12:38.916891 4730 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/691c1af1-e7b5-4b53-8f4d-9a33e48106be-registry-tls\") on node \"crc\" DevicePath \"\"" Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.044409 4730 generic.go:334] "Generic (PLEG): container finished" podID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" containerID="fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd" exitCode=0 Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.044460 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" event={"ID":"691c1af1-e7b5-4b53-8f4d-9a33e48106be","Type":"ContainerDied","Data":"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd"} Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.044489 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" event={"ID":"691c1af1-e7b5-4b53-8f4d-9a33e48106be","Type":"ContainerDied","Data":"97a7852ad22c32f44bbc120d697bfcdbaa504a08dec964c49cc1a7b55f2aed02"} Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.044507 4730 scope.go:117] "RemoveContainer" containerID="fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd" Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.044510 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-nnhll" Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.072107 4730 scope.go:117] "RemoveContainer" containerID="fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd" Feb 21 00:12:39 crc kubenswrapper[4730]: E0221 00:12:39.073310 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd\": container with ID starting with fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd not found: ID does not exist" containerID="fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd" Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.073368 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd"} err="failed to get container status \"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd\": rpc error: code = NotFound desc = could not find container \"fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd\": container with ID starting with fb7642b33c29ad2c6c074dea49667eb00e37524a403ee2b3974ec3361ff30ecd not found: ID does not exist" Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.093060 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:12:39 crc kubenswrapper[4730]: I0221 00:12:39.101303 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-nnhll"] Feb 21 00:12:40 crc kubenswrapper[4730]: I0221 00:12:40.449325 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" path="/var/lib/kubelet/pods/691c1af1-e7b5-4b53-8f4d-9a33e48106be/volumes" Feb 21 00:13:05 crc kubenswrapper[4730]: I0221 00:13:05.775244 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:13:05 crc kubenswrapper[4730]: I0221 00:13:05.775986 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:13:05 crc kubenswrapper[4730]: I0221 00:13:05.776055 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:13:05 crc kubenswrapper[4730]: I0221 00:13:05.777782 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587"} pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 21 00:13:05 crc kubenswrapper[4730]: I0221 00:13:05.777985 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" containerID="cri-o://b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587" gracePeriod=600 Feb 21 00:13:06 crc kubenswrapper[4730]: I0221 00:13:06.256235 4730 generic.go:334] "Generic (PLEG): container finished" podID="33513de3-5480-4aef-87ff-879f9e7a475f" containerID="b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587" exitCode=0 Feb 21 00:13:06 crc kubenswrapper[4730]: I0221 00:13:06.256457 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerDied","Data":"b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587"} Feb 21 00:13:06 crc kubenswrapper[4730]: I0221 00:13:06.256659 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c"} Feb 21 00:13:06 crc kubenswrapper[4730]: I0221 00:13:06.256690 4730 scope.go:117] "RemoveContainer" containerID="9ee78b3aeefd48d0281405b45bbb89f6465a707d704ebb48ff2e2419f2b5649c" Feb 21 00:14:48 crc kubenswrapper[4730]: I0221 00:14:48.755505 4730 scope.go:117] "RemoveContainer" containerID="3f694844b8366bd91f34a6a1fb0cc86b4182dd1145cac0b321e62387d5297478" Feb 21 00:14:48 crc kubenswrapper[4730]: I0221 00:14:48.777991 4730 scope.go:117] "RemoveContainer" containerID="15bdb7ede88d5b9da751da95e920fa2ddb139ffba22a0255d6705009aacd8f09" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.164663 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w"] Feb 21 00:15:00 crc kubenswrapper[4730]: E0221 00:15:00.165586 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" containerName="registry" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.165605 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" containerName="registry" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.165756 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="691c1af1-e7b5-4b53-8f4d-9a33e48106be" containerName="registry" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.166329 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.168866 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.168930 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w"] Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.169095 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.319468 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.319573 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7d4s\" (UniqueName: \"kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.319613 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.420561 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.420717 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7d4s\" (UniqueName: \"kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.420821 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.421874 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.431453 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.447891 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7d4s\" (UniqueName: \"kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s\") pod \"collect-profiles-29527215-d4q5w\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.489973 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:00 crc kubenswrapper[4730]: I0221 00:15:00.890432 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w"] Feb 21 00:15:01 crc kubenswrapper[4730]: I0221 00:15:01.666682 4730 generic.go:334] "Generic (PLEG): container finished" podID="86deba0d-565c-4d21-b040-d4e2b27bc65e" containerID="fa15aa934d36fdfd48e2f85a55d1742399d66a9b78ae606565faeb892d99e6d3" exitCode=0 Feb 21 00:15:01 crc kubenswrapper[4730]: I0221 00:15:01.666799 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" event={"ID":"86deba0d-565c-4d21-b040-d4e2b27bc65e","Type":"ContainerDied","Data":"fa15aa934d36fdfd48e2f85a55d1742399d66a9b78ae606565faeb892d99e6d3"} Feb 21 00:15:01 crc kubenswrapper[4730]: I0221 00:15:01.667190 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" event={"ID":"86deba0d-565c-4d21-b040-d4e2b27bc65e","Type":"ContainerStarted","Data":"a748a888ed47b3eb43102a10cd26858e22f688e1d378cd3602874d73c98ad461"} Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.871677 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.958029 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7d4s\" (UniqueName: \"kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s\") pod \"86deba0d-565c-4d21-b040-d4e2b27bc65e\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.958137 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume\") pod \"86deba0d-565c-4d21-b040-d4e2b27bc65e\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.958167 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume\") pod \"86deba0d-565c-4d21-b040-d4e2b27bc65e\" (UID: \"86deba0d-565c-4d21-b040-d4e2b27bc65e\") " Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.959092 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume" (OuterVolumeSpecName: "config-volume") pod "86deba0d-565c-4d21-b040-d4e2b27bc65e" (UID: "86deba0d-565c-4d21-b040-d4e2b27bc65e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.963265 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s" (OuterVolumeSpecName: "kube-api-access-s7d4s") pod "86deba0d-565c-4d21-b040-d4e2b27bc65e" (UID: "86deba0d-565c-4d21-b040-d4e2b27bc65e"). InnerVolumeSpecName "kube-api-access-s7d4s". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:15:02 crc kubenswrapper[4730]: I0221 00:15:02.963735 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "86deba0d-565c-4d21-b040-d4e2b27bc65e" (UID: "86deba0d-565c-4d21-b040-d4e2b27bc65e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.059596 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7d4s\" (UniqueName: \"kubernetes.io/projected/86deba0d-565c-4d21-b040-d4e2b27bc65e-kube-api-access-s7d4s\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.059651 4730 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/86deba0d-565c-4d21-b040-d4e2b27bc65e-secret-volume\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.059674 4730 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/86deba0d-565c-4d21-b040-d4e2b27bc65e-config-volume\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.682466 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.682377 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29527215-d4q5w" event={"ID":"86deba0d-565c-4d21-b040-d4e2b27bc65e","Type":"ContainerDied","Data":"a748a888ed47b3eb43102a10cd26858e22f688e1d378cd3602874d73c98ad461"} Feb 21 00:15:03 crc kubenswrapper[4730]: I0221 00:15:03.684266 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a748a888ed47b3eb43102a10cd26858e22f688e1d378cd3602874d73c98ad461" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.576571 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rjptr"] Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578326 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-controller" containerID="cri-o://b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578361 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="nbdb" containerID="cri-o://64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578416 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578490 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-node" containerID="cri-o://ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578533 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="sbdb" containerID="cri-o://902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578548 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="northd" containerID="cri-o://f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.578562 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-acl-logging" containerID="cri-o://510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.619201 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" containerID="cri-o://5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" gracePeriod=30 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.792466 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/2.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.793052 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/1.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.793204 4730 generic.go:334] "Generic (PLEG): container finished" podID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" containerID="ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432" exitCode=2 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.793265 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerDied","Data":"ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.793490 4730 scope.go:117] "RemoveContainer" containerID="ebaf1f6584d7b68999cbcd9353c761a721b83e0db59563503db35142cfd344c6" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.794157 4730 scope.go:117] "RemoveContainer" containerID="ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432" Feb 21 00:15:20 crc kubenswrapper[4730]: E0221 00:15:20.794533 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m9nnd_openshift-multus(ac51a893-4a0c-44ed-8284-0aac9e8d02ef)\"" pod="openshift-multus/multus-m9nnd" podUID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.798416 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovnkube-controller/3.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.809337 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-acl-logging/0.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.809975 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-controller/0.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810792 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" exitCode=0 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810812 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" exitCode=0 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810825 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" exitCode=0 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810835 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" exitCode=143 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810844 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" exitCode=143 Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810866 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810895 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810946 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810958 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.810970 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546"} Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.911534 4730 scope.go:117] "RemoveContainer" containerID="b7c0bb7ef39a7e8917e874b864d004082bdb069d21b66d39084926723fe0ca82" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.942238 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-acl-logging/0.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.943318 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-controller/0.log" Feb 21 00:15:20 crc kubenswrapper[4730]: I0221 00:15:20.944063 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.016975 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017052 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017088 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017130 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9jdnh\" (UniqueName: \"kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017171 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017202 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017229 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017259 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017301 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017333 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017365 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017400 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017438 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017473 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017508 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017560 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017591 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017621 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017657 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.017690 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch\") pod \"cfb5b15d-d281-4633-9930-905d05ed76e8\" (UID: \"cfb5b15d-d281-4633-9930-905d05ed76e8\") " Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.018011 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.018070 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.018106 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.018142 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.019688 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket" (OuterVolumeSpecName: "log-socket") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.019770 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.019809 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.019843 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.019874 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash" (OuterVolumeSpecName: "host-slash") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020478 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020533 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020568 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log" (OuterVolumeSpecName: "node-log") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020735 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020785 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.020862 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-cxf9s"] Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021097 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-node" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021121 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-node" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021135 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-ovn-metrics" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021143 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-ovn-metrics" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021158 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021166 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021174 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021182 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021191 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="nbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021199 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="nbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021209 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021217 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021229 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021236 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021245 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kubecfg-setup" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021252 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kubecfg-setup" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021261 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="northd" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021270 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="northd" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021278 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86deba0d-565c-4d21-b040-d4e2b27bc65e" containerName="collect-profiles" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021285 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="86deba0d-565c-4d21-b040-d4e2b27bc65e" containerName="collect-profiles" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021293 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-acl-logging" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021301 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-acl-logging" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021312 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="sbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021319 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="sbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021327 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021335 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: E0221 00:15:21.021343 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021352 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021460 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021472 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="northd" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021481 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="86deba0d-565c-4d21-b040-d4e2b27bc65e" containerName="collect-profiles" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021489 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021498 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021505 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="sbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021514 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021522 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="nbdb" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021531 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-acl-logging" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021540 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-node" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021548 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovn-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021556 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="kube-rbac-proxy-ovn-metrics" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021739 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerName="ovnkube-controller" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.021112 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.022518 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.022682 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.040240 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.043022 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.044949 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh" (OuterVolumeSpecName: "kube-api-access-9jdnh") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "kube-api-access-9jdnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.060398 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "cfb5b15d-d281-4633-9930-905d05ed76e8" (UID: "cfb5b15d-d281-4633-9930-905d05ed76e8"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118475 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-netd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118524 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-kubelet\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118539 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-systemd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118626 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59pr6\" (UniqueName: \"kubernetes.io/projected/7a166ac7-4247-4b93-b410-083cea9c8e98-kube-api-access-59pr6\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118654 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118677 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-slash\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118693 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-env-overrides\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118748 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118765 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-var-lib-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118781 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-netns\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118815 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-systemd-units\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118839 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118852 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-log-socket\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118867 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-etc-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118881 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-config\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118923 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-ovn\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118946 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-bin\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118978 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-node-log\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.118997 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-script-lib\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119020 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a166ac7-4247-4b93-b410-083cea9c8e98-ovn-node-metrics-cert\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119074 4730 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-config\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119084 4730 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-ovn\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119093 4730 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-node-log\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119101 4730 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-log-socket\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119109 4730 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-netns\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119118 4730 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-netd\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119126 4730 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cfb5b15d-d281-4633-9930-905d05ed76e8-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119135 4730 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-env-overrides\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119143 4730 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cfb5b15d-d281-4633-9930-905d05ed76e8-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119150 4730 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-systemd\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119158 4730 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119165 4730 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119173 4730 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119183 4730 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-cni-bin\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119194 4730 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-systemd-units\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119207 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9jdnh\" (UniqueName: \"kubernetes.io/projected/cfb5b15d-d281-4633-9930-905d05ed76e8-kube-api-access-9jdnh\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119216 4730 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-kubelet\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119224 4730 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119232 4730 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-run-openvswitch\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.119263 4730 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cfb5b15d-d281-4633-9930-905d05ed76e8-host-slash\") on node \"crc\" DevicePath \"\"" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220163 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a166ac7-4247-4b93-b410-083cea9c8e98-ovn-node-metrics-cert\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220253 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-netd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220292 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-kubelet\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220324 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-systemd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220355 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59pr6\" (UniqueName: \"kubernetes.io/projected/7a166ac7-4247-4b93-b410-083cea9c8e98-kube-api-access-59pr6\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220390 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220423 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-env-overrides\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220427 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-kubelet\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220451 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-slash\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220446 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-systemd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220454 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-netd\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220505 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-slash\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220540 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220727 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220790 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220822 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-var-lib-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220884 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-netns\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220948 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-var-lib-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.220982 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-run-netns\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221020 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-systemd-units\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221054 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221084 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-systemd-units\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221081 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-log-socket\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221126 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-etc-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221129 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-log-socket\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221145 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-config\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221168 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-ovn\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221184 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-bin\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221205 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-node-log\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221210 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-etc-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221244 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-script-lib\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221217 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-openvswitch\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221275 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-host-cni-bin\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221344 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-run-ovn\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221367 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-env-overrides\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221388 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/7a166ac7-4247-4b93-b410-083cea9c8e98-node-log\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.221950 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-config\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.222643 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/7a166ac7-4247-4b93-b410-083cea9c8e98-ovnkube-script-lib\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.224723 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/7a166ac7-4247-4b93-b410-083cea9c8e98-ovn-node-metrics-cert\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.239967 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59pr6\" (UniqueName: \"kubernetes.io/projected/7a166ac7-4247-4b93-b410-083cea9c8e98-kube-api-access-59pr6\") pod \"ovnkube-node-cxf9s\" (UID: \"7a166ac7-4247-4b93-b410-083cea9c8e98\") " pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.365005 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.823954 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-acl-logging/0.log" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.824831 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-rjptr_cfb5b15d-d281-4633-9930-905d05ed76e8/ovn-controller/0.log" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825689 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" exitCode=0 Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825735 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" exitCode=0 Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825760 4730 generic.go:334] "Generic (PLEG): container finished" podID="cfb5b15d-d281-4633-9930-905d05ed76e8" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" exitCode=0 Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825828 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825855 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825891 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825937 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825956 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-rjptr" event={"ID":"cfb5b15d-d281-4633-9930-905d05ed76e8","Type":"ContainerDied","Data":"90db4f07cf055e32c79857260824754a4e4f6c0f7ae993e96b137c91cfd994e6"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.825965 4730 scope.go:117] "RemoveContainer" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.833595 4730 generic.go:334] "Generic (PLEG): container finished" podID="7a166ac7-4247-4b93-b410-083cea9c8e98" containerID="cccfa619b0b2b2c2a85f0395003e90d8ac460c6e7240ab7543a790100739f2c9" exitCode=0 Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.833651 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerDied","Data":"cccfa619b0b2b2c2a85f0395003e90d8ac460c6e7240ab7543a790100739f2c9"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.833744 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"4e82acb537dd3d4fded0f4cf2bc5532d456659abbdae6e40e5955e3fd9cf2132"} Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.836519 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/2.log" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.852341 4730 scope.go:117] "RemoveContainer" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.898880 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rjptr"] Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.900104 4730 scope.go:117] "RemoveContainer" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.901075 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-rjptr"] Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.937035 4730 scope.go:117] "RemoveContainer" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.953522 4730 scope.go:117] "RemoveContainer" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" Feb 21 00:15:21 crc kubenswrapper[4730]: I0221 00:15:21.977435 4730 scope.go:117] "RemoveContainer" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.004268 4730 scope.go:117] "RemoveContainer" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.018070 4730 scope.go:117] "RemoveContainer" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.030719 4730 scope.go:117] "RemoveContainer" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.042172 4730 scope.go:117] "RemoveContainer" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.042690 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": container with ID starting with 5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8 not found: ID does not exist" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.042723 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8"} err="failed to get container status \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": rpc error: code = NotFound desc = could not find container \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": container with ID starting with 5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.042750 4730 scope.go:117] "RemoveContainer" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.043069 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": container with ID starting with 902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d not found: ID does not exist" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.043175 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d"} err="failed to get container status \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": rpc error: code = NotFound desc = could not find container \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": container with ID starting with 902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.043263 4730 scope.go:117] "RemoveContainer" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.043526 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": container with ID starting with 64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d not found: ID does not exist" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.043622 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d"} err="failed to get container status \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": rpc error: code = NotFound desc = could not find container \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": container with ID starting with 64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.043706 4730 scope.go:117] "RemoveContainer" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.043993 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": container with ID starting with f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061 not found: ID does not exist" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.044088 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061"} err="failed to get container status \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": rpc error: code = NotFound desc = could not find container \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": container with ID starting with f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.044164 4730 scope.go:117] "RemoveContainer" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.044573 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": container with ID starting with a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260 not found: ID does not exist" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.044667 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260"} err="failed to get container status \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": rpc error: code = NotFound desc = could not find container \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": container with ID starting with a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.044746 4730 scope.go:117] "RemoveContainer" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.045000 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": container with ID starting with ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03 not found: ID does not exist" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.045095 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03"} err="failed to get container status \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": rpc error: code = NotFound desc = could not find container \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": container with ID starting with ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.045175 4730 scope.go:117] "RemoveContainer" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.045417 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": container with ID starting with 510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717 not found: ID does not exist" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.045501 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717"} err="failed to get container status \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": rpc error: code = NotFound desc = could not find container \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": container with ID starting with 510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.045564 4730 scope.go:117] "RemoveContainer" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.046117 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": container with ID starting with b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546 not found: ID does not exist" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.046207 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546"} err="failed to get container status \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": rpc error: code = NotFound desc = could not find container \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": container with ID starting with b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.046330 4730 scope.go:117] "RemoveContainer" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" Feb 21 00:15:22 crc kubenswrapper[4730]: E0221 00:15:22.046611 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": container with ID starting with 61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975 not found: ID does not exist" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.046742 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975"} err="failed to get container status \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": rpc error: code = NotFound desc = could not find container \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": container with ID starting with 61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.046895 4730 scope.go:117] "RemoveContainer" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047187 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8"} err="failed to get container status \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": rpc error: code = NotFound desc = could not find container \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": container with ID starting with 5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047315 4730 scope.go:117] "RemoveContainer" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047694 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d"} err="failed to get container status \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": rpc error: code = NotFound desc = could not find container \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": container with ID starting with 902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047712 4730 scope.go:117] "RemoveContainer" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047865 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d"} err="failed to get container status \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": rpc error: code = NotFound desc = could not find container \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": container with ID starting with 64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.047882 4730 scope.go:117] "RemoveContainer" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048157 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061"} err="failed to get container status \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": rpc error: code = NotFound desc = could not find container \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": container with ID starting with f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048178 4730 scope.go:117] "RemoveContainer" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048475 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260"} err="failed to get container status \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": rpc error: code = NotFound desc = could not find container \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": container with ID starting with a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048613 4730 scope.go:117] "RemoveContainer" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048842 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03"} err="failed to get container status \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": rpc error: code = NotFound desc = could not find container \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": container with ID starting with ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.048952 4730 scope.go:117] "RemoveContainer" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049247 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717"} err="failed to get container status \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": rpc error: code = NotFound desc = could not find container \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": container with ID starting with 510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049327 4730 scope.go:117] "RemoveContainer" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049683 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546"} err="failed to get container status \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": rpc error: code = NotFound desc = could not find container \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": container with ID starting with b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049704 4730 scope.go:117] "RemoveContainer" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049929 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975"} err="failed to get container status \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": rpc error: code = NotFound desc = could not find container \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": container with ID starting with 61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.049950 4730 scope.go:117] "RemoveContainer" containerID="5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.050192 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8"} err="failed to get container status \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": rpc error: code = NotFound desc = could not find container \"5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8\": container with ID starting with 5b39b82fa25498db5538a325ccd7ddebd3c6e75c80382a37cf36e296daf477a8 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.050295 4730 scope.go:117] "RemoveContainer" containerID="902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.050547 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d"} err="failed to get container status \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": rpc error: code = NotFound desc = could not find container \"902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d\": container with ID starting with 902b348af55cd092d0b97d694b783cf4b8a78e597b70591e9eabaad9b3c4067d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.050851 4730 scope.go:117] "RemoveContainer" containerID="64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.051216 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d"} err="failed to get container status \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": rpc error: code = NotFound desc = could not find container \"64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d\": container with ID starting with 64b2e8de2bbf0bf9f37e8396e6ee0a9f07dbf1ac25972c451a782010dd949a8d not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.051315 4730 scope.go:117] "RemoveContainer" containerID="f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.051603 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061"} err="failed to get container status \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": rpc error: code = NotFound desc = could not find container \"f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061\": container with ID starting with f39003b1544d22a37c9aeab2671381692a0cc891d69d2128cf4ec583e3707061 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.051692 4730 scope.go:117] "RemoveContainer" containerID="a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.051963 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260"} err="failed to get container status \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": rpc error: code = NotFound desc = could not find container \"a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260\": container with ID starting with a2dfb4627b9e68f9fe3846a625958e2db642ffbbe9bb7de50050c4b2d0a40260 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052049 4730 scope.go:117] "RemoveContainer" containerID="ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052287 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03"} err="failed to get container status \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": rpc error: code = NotFound desc = could not find container \"ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03\": container with ID starting with ebe9eabc53cdbad4fe841ca7514481dbfef55b7401002caac4a635b2292d4f03 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052373 4730 scope.go:117] "RemoveContainer" containerID="510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052604 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717"} err="failed to get container status \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": rpc error: code = NotFound desc = could not find container \"510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717\": container with ID starting with 510873af24419eaf0619187ca8ceb5eaf9bc5f0082518af408800a8caae59717 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052691 4730 scope.go:117] "RemoveContainer" containerID="b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052933 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546"} err="failed to get container status \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": rpc error: code = NotFound desc = could not find container \"b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546\": container with ID starting with b624f860d3d85d4a72c47aed62a6a280934f6ba34658eb14f28942a6e1d30546 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.052957 4730 scope.go:117] "RemoveContainer" containerID="61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.053158 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975"} err="failed to get container status \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": rpc error: code = NotFound desc = could not find container \"61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975\": container with ID starting with 61337caa41dd34886f1fce39edf6be330fad04d2db014f8b2f541432133e1975 not found: ID does not exist" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.446009 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb5b15d-d281-4633-9930-905d05ed76e8" path="/var/lib/kubelet/pods/cfb5b15d-d281-4633-9930-905d05ed76e8/volumes" Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.844671 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"0f4673e26aaa8195a35e8aaa851d5e82c2bc66bdc9746b6e0fff93384da6d49d"} Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.844972 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"a3fc1632d8010ba5c56c2f5db8d779201ddb969069224bb03a09d8a1c097149f"} Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.844988 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"62667413b1e7ce30f9aceb7013cbb92e1c789a0a7b253f82c3aacd00c7c3c257"} Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.845000 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"1f14b8e6c0bb4a7510008e9dff125d63c6cec9344489faba7fe29e4a1183ad01"} Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.845010 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"8760efada531f2b48d6bddb53951ce41a798c320dcdd3e8a01dc6fdc189f1a7a"} Feb 21 00:15:22 crc kubenswrapper[4730]: I0221 00:15:22.845019 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"a55b183177b4753894e00f7ba2ef212c784e2f3ea45e610bd6fda78361587c5c"} Feb 21 00:15:25 crc kubenswrapper[4730]: I0221 00:15:25.877926 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"567b632cf52a476c2db2b1d6dcbe30871f0c8b1dd34a4bd31f057d64244220f6"} Feb 21 00:15:27 crc kubenswrapper[4730]: I0221 00:15:27.893006 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" event={"ID":"7a166ac7-4247-4b93-b410-083cea9c8e98","Type":"ContainerStarted","Data":"19ba88de1a5cb5eb241d1f1b978cf77103912e9e2a17b675f69670a2ab3dac0d"} Feb 21 00:15:27 crc kubenswrapper[4730]: I0221 00:15:27.893405 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:27 crc kubenswrapper[4730]: I0221 00:15:27.893444 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:27 crc kubenswrapper[4730]: I0221 00:15:27.919251 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:27 crc kubenswrapper[4730]: I0221 00:15:27.922741 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" podStartSLOduration=7.92272487 podStartE2EDuration="7.92272487s" podCreationTimestamp="2026-02-21 00:15:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:15:27.920079444 +0000 UTC m=+580.210915174" watchObservedRunningTime="2026-02-21 00:15:27.92272487 +0000 UTC m=+580.213560600" Feb 21 00:15:28 crc kubenswrapper[4730]: I0221 00:15:28.899480 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:28 crc kubenswrapper[4730]: I0221 00:15:28.992540 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:15:35 crc kubenswrapper[4730]: I0221 00:15:35.435529 4730 scope.go:117] "RemoveContainer" containerID="ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432" Feb 21 00:15:35 crc kubenswrapper[4730]: E0221 00:15:35.436372 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m9nnd_openshift-multus(ac51a893-4a0c-44ed-8284-0aac9e8d02ef)\"" pod="openshift-multus/multus-m9nnd" podUID="ac51a893-4a0c-44ed-8284-0aac9e8d02ef" Feb 21 00:15:35 crc kubenswrapper[4730]: I0221 00:15:35.775435 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:15:35 crc kubenswrapper[4730]: I0221 00:15:35.775528 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:15:48 crc kubenswrapper[4730]: I0221 00:15:48.814417 4730 scope.go:117] "RemoveContainer" containerID="366327187acd3a992ff0dfc6cf56f86bf773471681a97f6d51be8bfe969f45b1" Feb 21 00:15:50 crc kubenswrapper[4730]: I0221 00:15:50.435264 4730 scope.go:117] "RemoveContainer" containerID="ce671da224d58c1d0657c500427ceaeecd3a827419802c9e20ebe4d8959c3432" Feb 21 00:15:51 crc kubenswrapper[4730]: I0221 00:15:51.050497 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m9nnd_ac51a893-4a0c-44ed-8284-0aac9e8d02ef/kube-multus/2.log" Feb 21 00:15:51 crc kubenswrapper[4730]: I0221 00:15:51.050836 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m9nnd" event={"ID":"ac51a893-4a0c-44ed-8284-0aac9e8d02ef","Type":"ContainerStarted","Data":"31aa9585d92cffdccb242be91d1aa4f695165175c9fc9b6fcffbdd242377c0cf"} Feb 21 00:15:51 crc kubenswrapper[4730]: I0221 00:15:51.394609 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-cxf9s" Feb 21 00:16:05 crc kubenswrapper[4730]: I0221 00:16:05.774793 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:16:05 crc kubenswrapper[4730]: I0221 00:16:05.775446 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.266267 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.267121 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8ghsj" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="registry-server" containerID="cri-o://a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185" gracePeriod=30 Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.676807 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.836440 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mzzq\" (UniqueName: \"kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq\") pod \"55765fd7-a04e-4509-9fc6-2633eb09341b\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.836671 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities\") pod \"55765fd7-a04e-4509-9fc6-2633eb09341b\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.836720 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content\") pod \"55765fd7-a04e-4509-9fc6-2633eb09341b\" (UID: \"55765fd7-a04e-4509-9fc6-2633eb09341b\") " Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.838851 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities" (OuterVolumeSpecName: "utilities") pod "55765fd7-a04e-4509-9fc6-2633eb09341b" (UID: "55765fd7-a04e-4509-9fc6-2633eb09341b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.847222 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq" (OuterVolumeSpecName: "kube-api-access-9mzzq") pod "55765fd7-a04e-4509-9fc6-2633eb09341b" (UID: "55765fd7-a04e-4509-9fc6-2633eb09341b"). InnerVolumeSpecName "kube-api-access-9mzzq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.865399 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "55765fd7-a04e-4509-9fc6-2633eb09341b" (UID: "55765fd7-a04e-4509-9fc6-2633eb09341b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.938992 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.939366 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/55765fd7-a04e-4509-9fc6-2633eb09341b-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:23 crc kubenswrapper[4730]: I0221 00:16:23.939546 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mzzq\" (UniqueName: \"kubernetes.io/projected/55765fd7-a04e-4509-9fc6-2633eb09341b-kube-api-access-9mzzq\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.270011 4730 generic.go:334] "Generic (PLEG): container finished" podID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerID="a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185" exitCode=0 Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.270068 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerDied","Data":"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185"} Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.270199 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8ghsj" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.270620 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8ghsj" event={"ID":"55765fd7-a04e-4509-9fc6-2633eb09341b","Type":"ContainerDied","Data":"1bf28d76776b06a993d4eaac574311f141f2b54c548eeb7173f07a7b61c94360"} Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.270634 4730 scope.go:117] "RemoveContainer" containerID="a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.289410 4730 scope.go:117] "RemoveContainer" containerID="81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.303634 4730 scope.go:117] "RemoveContainer" containerID="91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.333493 4730 scope.go:117] "RemoveContainer" containerID="a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185" Feb 21 00:16:24 crc kubenswrapper[4730]: E0221 00:16:24.333874 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185\": container with ID starting with a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185 not found: ID does not exist" containerID="a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.334044 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185"} err="failed to get container status \"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185\": rpc error: code = NotFound desc = could not find container \"a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185\": container with ID starting with a1d038dd3af8e2a12c10a2ddeef988b92c564b54cf9deca924a08da78583d185 not found: ID does not exist" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.334130 4730 scope.go:117] "RemoveContainer" containerID="81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.335243 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:16:24 crc kubenswrapper[4730]: E0221 00:16:24.335704 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b\": container with ID starting with 81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b not found: ID does not exist" containerID="81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.335745 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b"} err="failed to get container status \"81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b\": rpc error: code = NotFound desc = could not find container \"81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b\": container with ID starting with 81c4619cfd8e6a6686c0a4bbe19693fb831f2114a70a96aa1b1e375729cea51b not found: ID does not exist" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.335784 4730 scope.go:117] "RemoveContainer" containerID="91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae" Feb 21 00:16:24 crc kubenswrapper[4730]: E0221 00:16:24.336221 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae\": container with ID starting with 91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae not found: ID does not exist" containerID="91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.336305 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae"} err="failed to get container status \"91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae\": rpc error: code = NotFound desc = could not find container \"91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae\": container with ID starting with 91e785a0380da7f4857cd0026cbfc349ddc1ab0f85747af396b14f2739c9cdae not found: ID does not exist" Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.340410 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8ghsj"] Feb 21 00:16:24 crc kubenswrapper[4730]: I0221 00:16:24.447536 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" path="/var/lib/kubelet/pods/55765fd7-a04e-4509-9fc6-2633eb09341b/volumes" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.011595 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk"] Feb 21 00:16:27 crc kubenswrapper[4730]: E0221 00:16:27.012162 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="registry-server" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.012177 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="registry-server" Feb 21 00:16:27 crc kubenswrapper[4730]: E0221 00:16:27.012195 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="extract-content" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.012206 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="extract-content" Feb 21 00:16:27 crc kubenswrapper[4730]: E0221 00:16:27.012230 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="extract-utilities" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.012240 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="extract-utilities" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.012367 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="55765fd7-a04e-4509-9fc6-2633eb09341b" containerName="registry-server" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.013228 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.015994 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.021033 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk"] Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.184180 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fk4rt\" (UniqueName: \"kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.184688 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.184838 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.286131 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.286582 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.286675 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fk4rt\" (UniqueName: \"kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.287066 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.287073 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.309789 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fk4rt\" (UniqueName: \"kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.370418 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:27 crc kubenswrapper[4730]: I0221 00:16:27.829204 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk"] Feb 21 00:16:28 crc kubenswrapper[4730]: I0221 00:16:28.297965 4730 generic.go:334] "Generic (PLEG): container finished" podID="019f59b9-69c1-412c-91e7-38539454c642" containerID="04846745cd64e68e2180032eac053b0cb2910dcd4e27149e7cbe4236c262b35a" exitCode=0 Feb 21 00:16:28 crc kubenswrapper[4730]: I0221 00:16:28.298018 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" event={"ID":"019f59b9-69c1-412c-91e7-38539454c642","Type":"ContainerDied","Data":"04846745cd64e68e2180032eac053b0cb2910dcd4e27149e7cbe4236c262b35a"} Feb 21 00:16:28 crc kubenswrapper[4730]: I0221 00:16:28.298274 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" event={"ID":"019f59b9-69c1-412c-91e7-38539454c642","Type":"ContainerStarted","Data":"daefd47addd6975601a3b5cc95c0e5839cfe60145ad35d19b1f7c242b3a4a89f"} Feb 21 00:16:28 crc kubenswrapper[4730]: I0221 00:16:28.301205 4730 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 21 00:16:30 crc kubenswrapper[4730]: I0221 00:16:30.312247 4730 generic.go:334] "Generic (PLEG): container finished" podID="019f59b9-69c1-412c-91e7-38539454c642" containerID="ff8a8076174a077fe7da23d3eccd33123f15dd670bfbecc5efd9ae3d83994113" exitCode=0 Feb 21 00:16:30 crc kubenswrapper[4730]: I0221 00:16:30.312360 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" event={"ID":"019f59b9-69c1-412c-91e7-38539454c642","Type":"ContainerDied","Data":"ff8a8076174a077fe7da23d3eccd33123f15dd670bfbecc5efd9ae3d83994113"} Feb 21 00:16:31 crc kubenswrapper[4730]: I0221 00:16:31.323106 4730 generic.go:334] "Generic (PLEG): container finished" podID="019f59b9-69c1-412c-91e7-38539454c642" containerID="17dabbb62cd4bfb0c1a4e1c3be5844c46a0359ef196c1e6d1081811f2311d57c" exitCode=0 Feb 21 00:16:31 crc kubenswrapper[4730]: I0221 00:16:31.323165 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" event={"ID":"019f59b9-69c1-412c-91e7-38539454c642","Type":"ContainerDied","Data":"17dabbb62cd4bfb0c1a4e1c3be5844c46a0359ef196c1e6d1081811f2311d57c"} Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.677503 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.877392 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle\") pod \"019f59b9-69c1-412c-91e7-38539454c642\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.877551 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fk4rt\" (UniqueName: \"kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt\") pod \"019f59b9-69c1-412c-91e7-38539454c642\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.877595 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util\") pod \"019f59b9-69c1-412c-91e7-38539454c642\" (UID: \"019f59b9-69c1-412c-91e7-38539454c642\") " Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.881263 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle" (OuterVolumeSpecName: "bundle") pod "019f59b9-69c1-412c-91e7-38539454c642" (UID: "019f59b9-69c1-412c-91e7-38539454c642"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.888068 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt" (OuterVolumeSpecName: "kube-api-access-fk4rt") pod "019f59b9-69c1-412c-91e7-38539454c642" (UID: "019f59b9-69c1-412c-91e7-38539454c642"). InnerVolumeSpecName "kube-api-access-fk4rt". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.898029 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util" (OuterVolumeSpecName: "util") pod "019f59b9-69c1-412c-91e7-38539454c642" (UID: "019f59b9-69c1-412c-91e7-38539454c642"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.979558 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fk4rt\" (UniqueName: \"kubernetes.io/projected/019f59b9-69c1-412c-91e7-38539454c642-kube-api-access-fk4rt\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.979611 4730 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-util\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:32 crc kubenswrapper[4730]: I0221 00:16:32.979632 4730 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/019f59b9-69c1-412c-91e7-38539454c642-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.339692 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" event={"ID":"019f59b9-69c1-412c-91e7-38539454c642","Type":"ContainerDied","Data":"daefd47addd6975601a3b5cc95c0e5839cfe60145ad35d19b1f7c242b3a4a89f"} Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.339989 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="daefd47addd6975601a3b5cc95c0e5839cfe60145ad35d19b1f7c242b3a4a89f" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.339870 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.633463 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4"] Feb 21 00:16:33 crc kubenswrapper[4730]: E0221 00:16:33.633928 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="extract" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.633949 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="extract" Feb 21 00:16:33 crc kubenswrapper[4730]: E0221 00:16:33.633977 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="util" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.633992 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="util" Feb 21 00:16:33 crc kubenswrapper[4730]: E0221 00:16:33.634017 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="pull" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.634070 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="pull" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.634458 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="019f59b9-69c1-412c-91e7-38539454c642" containerName="extract" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.636521 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.642896 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.650696 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4"] Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.694691 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.694748 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnf72\" (UniqueName: \"kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.694780 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.795674 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.795759 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.795786 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnf72\" (UniqueName: \"kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.796318 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.796525 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.823623 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnf72\" (UniqueName: \"kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:33 crc kubenswrapper[4730]: I0221 00:16:33.995313 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.260536 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4"] Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.346378 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerStarted","Data":"881234b0fe11f225ccdf8e3cf646a6078134445648e0141987089a0909750bf9"} Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.414261 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz"] Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.415976 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.424685 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz"] Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.605996 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.606083 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49fjf\" (UniqueName: \"kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.606173 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.706977 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.707072 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49fjf\" (UniqueName: \"kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.707161 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.707761 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.707892 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:34 crc kubenswrapper[4730]: I0221 00:16:34.740125 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49fjf\" (UniqueName: \"kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf\") pod \"00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.037189 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.320434 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz"] Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.353554 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerStarted","Data":"2ad75f6e39bab62d3d8bd5d2fed7d1ab60122dbbc250cdb62ad9ed28fdd50115"} Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.355080 4730 generic.go:334] "Generic (PLEG): container finished" podID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerID="9d0e7ad6ccc1fa1f9a4078e4f8b63023a4087d3241a2d794ff4b029d0ec5617c" exitCode=0 Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.355402 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerDied","Data":"9d0e7ad6ccc1fa1f9a4078e4f8b63023a4087d3241a2d794ff4b029d0ec5617c"} Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.775003 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.775612 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.775699 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.776850 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c"} pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 21 00:16:35 crc kubenswrapper[4730]: I0221 00:16:35.777042 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" containerID="cri-o://7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c" gracePeriod=600 Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.362540 4730 generic.go:334] "Generic (PLEG): container finished" podID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerID="f0a4445a89d4f09d9cfe54d7820cd1c39fdf4f94b4743ca2174dd27ff044d248" exitCode=0 Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.362589 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerDied","Data":"f0a4445a89d4f09d9cfe54d7820cd1c39fdf4f94b4743ca2174dd27ff044d248"} Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.366408 4730 generic.go:334] "Generic (PLEG): container finished" podID="33513de3-5480-4aef-87ff-879f9e7a475f" containerID="7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c" exitCode=0 Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.366467 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerDied","Data":"7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c"} Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.366525 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e"} Feb 21 00:16:36 crc kubenswrapper[4730]: I0221 00:16:36.366554 4730 scope.go:117] "RemoveContainer" containerID="b14b7b520f0e811f4b4c0cea5dd884aef27b179ef299cd78811117075b475587" Feb 21 00:16:37 crc kubenswrapper[4730]: I0221 00:16:37.377839 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerStarted","Data":"17e6cc8e84e9720c26a943f611c517eeeedaeb094880b818f549f99174f3515e"} Feb 21 00:16:37 crc kubenswrapper[4730]: I0221 00:16:37.379538 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerStarted","Data":"d4bcfb265689d58ce02b538bfdb50afa44cf72620ac92b261362895ae3235dec"} Feb 21 00:16:38 crc kubenswrapper[4730]: I0221 00:16:38.390018 4730 generic.go:334] "Generic (PLEG): container finished" podID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerID="d4bcfb265689d58ce02b538bfdb50afa44cf72620ac92b261362895ae3235dec" exitCode=0 Feb 21 00:16:38 crc kubenswrapper[4730]: I0221 00:16:38.390076 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerDied","Data":"d4bcfb265689d58ce02b538bfdb50afa44cf72620ac92b261362895ae3235dec"} Feb 21 00:16:38 crc kubenswrapper[4730]: I0221 00:16:38.393077 4730 generic.go:334] "Generic (PLEG): container finished" podID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerID="17e6cc8e84e9720c26a943f611c517eeeedaeb094880b818f549f99174f3515e" exitCode=0 Feb 21 00:16:38 crc kubenswrapper[4730]: I0221 00:16:38.393127 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerDied","Data":"17e6cc8e84e9720c26a943f611c517eeeedaeb094880b818f549f99174f3515e"} Feb 21 00:16:39 crc kubenswrapper[4730]: I0221 00:16:39.400726 4730 generic.go:334] "Generic (PLEG): container finished" podID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerID="48cb4a21eccb8debb648eb0b0dc878d75811e0629736360454f1f77e0fb04648" exitCode=0 Feb 21 00:16:39 crc kubenswrapper[4730]: I0221 00:16:39.400768 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerDied","Data":"48cb4a21eccb8debb648eb0b0dc878d75811e0629736360454f1f77e0fb04648"} Feb 21 00:16:39 crc kubenswrapper[4730]: I0221 00:16:39.403014 4730 generic.go:334] "Generic (PLEG): container finished" podID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerID="bf428bfd3ecbef0d139455446a514d3d7551031e44d2807b9bf6b443e3875b84" exitCode=0 Feb 21 00:16:39 crc kubenswrapper[4730]: I0221 00:16:39.403057 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerDied","Data":"bf428bfd3ecbef0d139455446a514d3d7551031e44d2807b9bf6b443e3875b84"} Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.794321 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.828404 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.893546 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util\") pod \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.893636 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle\") pod \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.893657 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49fjf\" (UniqueName: \"kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf\") pod \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\" (UID: \"e8441d0d-ff6d-4b25-9460-cf08a6f099dd\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.895359 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle" (OuterVolumeSpecName: "bundle") pod "e8441d0d-ff6d-4b25-9460-cf08a6f099dd" (UID: "e8441d0d-ff6d-4b25-9460-cf08a6f099dd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.899613 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf" (OuterVolumeSpecName: "kube-api-access-49fjf") pod "e8441d0d-ff6d-4b25-9460-cf08a6f099dd" (UID: "e8441d0d-ff6d-4b25-9460-cf08a6f099dd"). InnerVolumeSpecName "kube-api-access-49fjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.958574 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util" (OuterVolumeSpecName: "util") pod "e8441d0d-ff6d-4b25-9460-cf08a6f099dd" (UID: "e8441d0d-ff6d-4b25-9460-cf08a6f099dd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.994833 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util\") pod \"adc2b228-59a9-41d8-a948-83addaa4b8f8\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.994967 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnf72\" (UniqueName: \"kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72\") pod \"adc2b228-59a9-41d8-a948-83addaa4b8f8\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995017 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle\") pod \"adc2b228-59a9-41d8-a948-83addaa4b8f8\" (UID: \"adc2b228-59a9-41d8-a948-83addaa4b8f8\") " Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995462 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle" (OuterVolumeSpecName: "bundle") pod "adc2b228-59a9-41d8-a948-83addaa4b8f8" (UID: "adc2b228-59a9-41d8-a948-83addaa4b8f8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995661 4730 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-util\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995678 4730 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995687 4730 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:40 crc kubenswrapper[4730]: I0221 00:16:40.995697 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49fjf\" (UniqueName: \"kubernetes.io/projected/e8441d0d-ff6d-4b25-9460-cf08a6f099dd-kube-api-access-49fjf\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.007423 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72" (OuterVolumeSpecName: "kube-api-access-qnf72") pod "adc2b228-59a9-41d8-a948-83addaa4b8f8" (UID: "adc2b228-59a9-41d8-a948-83addaa4b8f8"). InnerVolumeSpecName "kube-api-access-qnf72". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.012895 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util" (OuterVolumeSpecName: "util") pod "adc2b228-59a9-41d8-a948-83addaa4b8f8" (UID: "adc2b228-59a9-41d8-a948-83addaa4b8f8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.097010 4730 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/adc2b228-59a9-41d8-a948-83addaa4b8f8-util\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.097047 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnf72\" (UniqueName: \"kubernetes.io/projected/adc2b228-59a9-41d8-a948-83addaa4b8f8-kube-api-access-qnf72\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.415370 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" event={"ID":"e8441d0d-ff6d-4b25-9460-cf08a6f099dd","Type":"ContainerDied","Data":"2ad75f6e39bab62d3d8bd5d2fed7d1ab60122dbbc250cdb62ad9ed28fdd50115"} Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.415412 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ad75f6e39bab62d3d8bd5d2fed7d1ab60122dbbc250cdb62ad9ed28fdd50115" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.415476 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.431466 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" event={"ID":"adc2b228-59a9-41d8-a948-83addaa4b8f8","Type":"ContainerDied","Data":"881234b0fe11f225ccdf8e3cf646a6078134445648e0141987089a0909750bf9"} Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.431495 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="881234b0fe11f225ccdf8e3cf646a6078134445648e0141987089a0909750bf9" Feb 21 00:16:41 crc kubenswrapper[4730]: I0221 00:16:41.431554 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719219 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss"] Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719452 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="pull" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719467 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="pull" Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719479 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719485 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719494 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="util" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719502 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="util" Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719517 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719526 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719538 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="util" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719547 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="util" Feb 21 00:16:42 crc kubenswrapper[4730]: E0221 00:16:42.719555 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="pull" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719562 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="pull" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719681 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8441d0d-ff6d-4b25-9460-cf08a6f099dd" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.719694 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="adc2b228-59a9-41d8-a948-83addaa4b8f8" containerName="extract" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.720406 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.723494 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.743059 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss"] Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.815785 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.815868 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c4fkb\" (UniqueName: \"kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.815936 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.916956 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.917024 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.917087 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c4fkb\" (UniqueName: \"kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.917715 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.917769 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:42 crc kubenswrapper[4730]: I0221 00:16:42.939255 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c4fkb\" (UniqueName: \"kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:43 crc kubenswrapper[4730]: I0221 00:16:43.036504 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:43 crc kubenswrapper[4730]: I0221 00:16:43.516115 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss"] Feb 21 00:16:43 crc kubenswrapper[4730]: W0221 00:16:43.527156 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod64321551_df8b_4344_a637_b21a8ab305ea.slice/crio-63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f WatchSource:0}: Error finding container 63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f: Status 404 returned error can't find the container with id 63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f Feb 21 00:16:44 crc kubenswrapper[4730]: I0221 00:16:44.445755 4730 generic.go:334] "Generic (PLEG): container finished" podID="64321551-df8b-4344-a637-b21a8ab305ea" containerID="50ad97324f7b182513d902f84b9f643efef0dd767723101ccf825755af78d4c6" exitCode=0 Feb 21 00:16:44 crc kubenswrapper[4730]: I0221 00:16:44.445992 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" event={"ID":"64321551-df8b-4344-a637-b21a8ab305ea","Type":"ContainerDied","Data":"50ad97324f7b182513d902f84b9f643efef0dd767723101ccf825755af78d4c6"} Feb 21 00:16:44 crc kubenswrapper[4730]: I0221 00:16:44.446076 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" event={"ID":"64321551-df8b-4344-a637-b21a8ab305ea","Type":"ContainerStarted","Data":"63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f"} Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.572005 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.573215 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.574863 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.576067 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-xprmn" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.576133 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.584386 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.676510 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.677342 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.681177 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.681424 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-s4hz9" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.690939 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.696572 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.697184 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.709819 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.753536 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2wtt\" (UniqueName: \"kubernetes.io/projected/018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd-kube-api-access-r2wtt\") pod \"obo-prometheus-operator-68bc856cb9-kfkqw\" (UID: \"018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.785652 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-wh7ls"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.786542 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.792248 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-pr7dg" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.792424 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.801149 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-wh7ls"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.855363 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.855503 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.855582 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.855634 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2wtt\" (UniqueName: \"kubernetes.io/projected/018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd-kube-api-access-r2wtt\") pod \"obo-prometheus-operator-68bc856cb9-kfkqw\" (UID: \"018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.855670 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.892801 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-9vq9k"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.894002 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.898379 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-wzbqm" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.911302 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2wtt\" (UniqueName: \"kubernetes.io/projected/018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd-kube-api-access-r2wtt\") pod \"obo-prometheus-operator-68bc856cb9-kfkqw\" (UID: \"018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.919683 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-9vq9k"] Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.956513 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrw62\" (UniqueName: \"kubernetes.io/projected/e7055108-18d7-4bc4-8644-a8d80b1605b3-kube-api-access-jrw62\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.956797 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.956933 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7055108-18d7-4bc4-8644-a8d80b1605b3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.957057 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.957158 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.957246 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.960758 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.960921 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5e6832e1-eb7d-4f11-9e11-95c61cbf0934-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7\" (UID: \"5e6832e1-eb7d-4f11-9e11-95c61cbf0934\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.962666 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.965319 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7d174177-75da-4fb4-a195-a3e06804ba1e-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5\" (UID: \"7d174177-75da-4fb4-a195-a3e06804ba1e\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:45 crc kubenswrapper[4730]: I0221 00:16:45.992947 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.009361 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.058130 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrw62\" (UniqueName: \"kubernetes.io/projected/e7055108-18d7-4bc4-8644-a8d80b1605b3-kube-api-access-jrw62\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.058202 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b9cd1a14-f993-46fc-816b-884442179c47-openshift-service-ca\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.058230 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgp7d\" (UniqueName: \"kubernetes.io/projected/b9cd1a14-f993-46fc-816b-884442179c47-kube-api-access-pgp7d\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.058261 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7055108-18d7-4bc4-8644-a8d80b1605b3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.064630 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/e7055108-18d7-4bc4-8644-a8d80b1605b3-observability-operator-tls\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.090427 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrw62\" (UniqueName: \"kubernetes.io/projected/e7055108-18d7-4bc4-8644-a8d80b1605b3-kube-api-access-jrw62\") pod \"observability-operator-59bdc8b94-wh7ls\" (UID: \"e7055108-18d7-4bc4-8644-a8d80b1605b3\") " pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.104594 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.165633 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b9cd1a14-f993-46fc-816b-884442179c47-openshift-service-ca\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.165960 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgp7d\" (UniqueName: \"kubernetes.io/projected/b9cd1a14-f993-46fc-816b-884442179c47-kube-api-access-pgp7d\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.166754 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/b9cd1a14-f993-46fc-816b-884442179c47-openshift-service-ca\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.189838 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgp7d\" (UniqueName: \"kubernetes.io/projected/b9cd1a14-f993-46fc-816b-884442179c47-kube-api-access-pgp7d\") pod \"perses-operator-5bf474d74f-9vq9k\" (UID: \"b9cd1a14-f993-46fc-816b-884442179c47\") " pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.190352 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.237695 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.343054 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7"] Feb 21 00:16:46 crc kubenswrapper[4730]: W0221 00:16:46.348098 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e6832e1_eb7d_4f11_9e11_95c61cbf0934.slice/crio-62d7f6ac8b4742af7c7dd87ad36dde719d8170d1db972e826e5911b0507dc3d0 WatchSource:0}: Error finding container 62d7f6ac8b4742af7c7dd87ad36dde719d8170d1db972e826e5911b0507dc3d0: Status 404 returned error can't find the container with id 62d7f6ac8b4742af7c7dd87ad36dde719d8170d1db972e826e5911b0507dc3d0 Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.484949 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" event={"ID":"5e6832e1-eb7d-4f11-9e11-95c61cbf0934","Type":"ContainerStarted","Data":"62d7f6ac8b4742af7c7dd87ad36dde719d8170d1db972e826e5911b0507dc3d0"} Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.515411 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw"] Feb 21 00:16:46 crc kubenswrapper[4730]: W0221 00:16:46.531301 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod018ca2e3_2fff_4b8c_b5b1_3d56c9d3ebfd.slice/crio-3d52a520e51d1aaec5452618c5ae97fac510d70143425a5be4a6d5245671d8fc WatchSource:0}: Error finding container 3d52a520e51d1aaec5452618c5ae97fac510d70143425a5be4a6d5245671d8fc: Status 404 returned error can't find the container with id 3d52a520e51d1aaec5452618c5ae97fac510d70143425a5be4a6d5245671d8fc Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.542451 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5"] Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.574029 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-9vq9k"] Feb 21 00:16:46 crc kubenswrapper[4730]: W0221 00:16:46.586197 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9cd1a14_f993_46fc_816b_884442179c47.slice/crio-61c6580b5f56aeb87ba67494d49d65b0baf9ec1d934cb2c9f8e3157b8ba2fbd2 WatchSource:0}: Error finding container 61c6580b5f56aeb87ba67494d49d65b0baf9ec1d934cb2c9f8e3157b8ba2fbd2: Status 404 returned error can't find the container with id 61c6580b5f56aeb87ba67494d49d65b0baf9ec1d934cb2c9f8e3157b8ba2fbd2 Feb 21 00:16:46 crc kubenswrapper[4730]: I0221 00:16:46.667646 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-wh7ls"] Feb 21 00:16:46 crc kubenswrapper[4730]: W0221 00:16:46.672937 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7055108_18d7_4bc4_8644_a8d80b1605b3.slice/crio-785b6875dc374dec6b58a162a76e9a326f347899f24da56379116bb55a95ddc9 WatchSource:0}: Error finding container 785b6875dc374dec6b58a162a76e9a326f347899f24da56379116bb55a95ddc9: Status 404 returned error can't find the container with id 785b6875dc374dec6b58a162a76e9a326f347899f24da56379116bb55a95ddc9 Feb 21 00:16:47 crc kubenswrapper[4730]: I0221 00:16:47.496296 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" event={"ID":"7d174177-75da-4fb4-a195-a3e06804ba1e","Type":"ContainerStarted","Data":"62afcc2d3bf9378002689470716c341bb3eee30297e11037922009eca80f8d77"} Feb 21 00:16:47 crc kubenswrapper[4730]: I0221 00:16:47.497179 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" event={"ID":"b9cd1a14-f993-46fc-816b-884442179c47","Type":"ContainerStarted","Data":"61c6580b5f56aeb87ba67494d49d65b0baf9ec1d934cb2c9f8e3157b8ba2fbd2"} Feb 21 00:16:47 crc kubenswrapper[4730]: I0221 00:16:47.497954 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" event={"ID":"e7055108-18d7-4bc4-8644-a8d80b1605b3","Type":"ContainerStarted","Data":"785b6875dc374dec6b58a162a76e9a326f347899f24da56379116bb55a95ddc9"} Feb 21 00:16:47 crc kubenswrapper[4730]: I0221 00:16:47.498668 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" event={"ID":"018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd","Type":"ContainerStarted","Data":"3d52a520e51d1aaec5452618c5ae97fac510d70143425a5be4a6d5245671d8fc"} Feb 21 00:16:48 crc kubenswrapper[4730]: I0221 00:16:48.858255 4730 scope.go:117] "RemoveContainer" containerID="37c6eb74e229c8794c48181f96eb406807a6a6bccbdd1e28a91cbe38147a06eb" Feb 21 00:16:49 crc kubenswrapper[4730]: I0221 00:16:49.517018 4730 generic.go:334] "Generic (PLEG): container finished" podID="64321551-df8b-4344-a637-b21a8ab305ea" containerID="aa97ad644976a76714724a475e2740cb71416b64ab680c764a7825353709ca88" exitCode=0 Feb 21 00:16:49 crc kubenswrapper[4730]: I0221 00:16:49.517129 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" event={"ID":"64321551-df8b-4344-a637-b21a8ab305ea","Type":"ContainerDied","Data":"aa97ad644976a76714724a475e2740cb71416b64ab680c764a7825353709ca88"} Feb 21 00:16:50 crc kubenswrapper[4730]: I0221 00:16:50.526448 4730 generic.go:334] "Generic (PLEG): container finished" podID="64321551-df8b-4344-a637-b21a8ab305ea" containerID="ee4bc0ac7b6784887c7e07d5227fe2e3b9b91031424a5c3f7f3fc379058fd60d" exitCode=0 Feb 21 00:16:50 crc kubenswrapper[4730]: I0221 00:16:50.526488 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" event={"ID":"64321551-df8b-4344-a637-b21a8ab305ea","Type":"ContainerDied","Data":"ee4bc0ac7b6784887c7e07d5227fe2e3b9b91031424a5c3f7f3fc379058fd60d"} Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.845755 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-7wv8x"] Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.846845 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.849845 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.851242 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.855468 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-7wv8x"] Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.858383 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-fndnh" Feb 21 00:16:51 crc kubenswrapper[4730]: I0221 00:16:51.952432 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwmmq\" (UniqueName: \"kubernetes.io/projected/70369d12-f9ef-4028-9a45-62aad9cca5e5-kube-api-access-fwmmq\") pod \"interconnect-operator-5bb49f789d-7wv8x\" (UID: \"70369d12-f9ef-4028-9a45-62aad9cca5e5\") " pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" Feb 21 00:16:52 crc kubenswrapper[4730]: I0221 00:16:52.053480 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwmmq\" (UniqueName: \"kubernetes.io/projected/70369d12-f9ef-4028-9a45-62aad9cca5e5-kube-api-access-fwmmq\") pod \"interconnect-operator-5bb49f789d-7wv8x\" (UID: \"70369d12-f9ef-4028-9a45-62aad9cca5e5\") " pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" Feb 21 00:16:52 crc kubenswrapper[4730]: I0221 00:16:52.076994 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwmmq\" (UniqueName: \"kubernetes.io/projected/70369d12-f9ef-4028-9a45-62aad9cca5e5-kube-api-access-fwmmq\") pod \"interconnect-operator-5bb49f789d-7wv8x\" (UID: \"70369d12-f9ef-4028-9a45-62aad9cca5e5\") " pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" Feb 21 00:16:52 crc kubenswrapper[4730]: I0221 00:16:52.168511 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.383711 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.570078 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c4fkb\" (UniqueName: \"kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb\") pod \"64321551-df8b-4344-a637-b21a8ab305ea\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.570373 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle\") pod \"64321551-df8b-4344-a637-b21a8ab305ea\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.570432 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util\") pod \"64321551-df8b-4344-a637-b21a8ab305ea\" (UID: \"64321551-df8b-4344-a637-b21a8ab305ea\") " Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.572509 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle" (OuterVolumeSpecName: "bundle") pod "64321551-df8b-4344-a637-b21a8ab305ea" (UID: "64321551-df8b-4344-a637-b21a8ab305ea"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.579507 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb" (OuterVolumeSpecName: "kube-api-access-c4fkb") pod "64321551-df8b-4344-a637-b21a8ab305ea" (UID: "64321551-df8b-4344-a637-b21a8ab305ea"). InnerVolumeSpecName "kube-api-access-c4fkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.581919 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util" (OuterVolumeSpecName: "util") pod "64321551-df8b-4344-a637-b21a8ab305ea" (UID: "64321551-df8b-4344-a637-b21a8ab305ea"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.594063 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" event={"ID":"64321551-df8b-4344-a637-b21a8ab305ea","Type":"ContainerDied","Data":"63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f"} Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.594097 4730 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="63ad76126a19a3896bbb430bcd28c26f0ce5fe65f8cbcfb5aae4202cb278c66f" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.594156 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.671575 4730 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-bundle\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.671616 4730 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/64321551-df8b-4344-a637-b21a8ab305ea-util\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:53 crc kubenswrapper[4730]: I0221 00:16:53.671631 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c4fkb\" (UniqueName: \"kubernetes.io/projected/64321551-df8b-4344-a637-b21a8ab305ea-kube-api-access-c4fkb\") on node \"crc\" DevicePath \"\"" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.675807 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-658c75dcc7-mnxjh"] Feb 21 00:16:54 crc kubenswrapper[4730]: E0221 00:16:54.676130 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="util" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.676150 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="util" Feb 21 00:16:54 crc kubenswrapper[4730]: E0221 00:16:54.676168 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="extract" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.676180 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="extract" Feb 21 00:16:54 crc kubenswrapper[4730]: E0221 00:16:54.676199 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="pull" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.676212 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="pull" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.676377 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="64321551-df8b-4344-a637-b21a8ab305ea" containerName="extract" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.677002 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.679460 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.679631 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-h6gzw" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.701042 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-658c75dcc7-mnxjh"] Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.788181 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hphwk\" (UniqueName: \"kubernetes.io/projected/158c1ae3-ea68-47d1-bc9e-654f1b566517-kube-api-access-hphwk\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.788562 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-webhook-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.788743 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-apiservice-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.889740 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hphwk\" (UniqueName: \"kubernetes.io/projected/158c1ae3-ea68-47d1-bc9e-654f1b566517-kube-api-access-hphwk\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.889787 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-webhook-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.889809 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-apiservice-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.893291 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-apiservice-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.893482 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/158c1ae3-ea68-47d1-bc9e-654f1b566517-webhook-cert\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.904193 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hphwk\" (UniqueName: \"kubernetes.io/projected/158c1ae3-ea68-47d1-bc9e-654f1b566517-kube-api-access-hphwk\") pod \"elastic-operator-658c75dcc7-mnxjh\" (UID: \"158c1ae3-ea68-47d1-bc9e-654f1b566517\") " pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:54 crc kubenswrapper[4730]: I0221 00:16:54.990872 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.277759 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-7wv8x"] Feb 21 00:16:58 crc kubenswrapper[4730]: W0221 00:16:58.340633 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70369d12_f9ef_4028_9a45_62aad9cca5e5.slice/crio-37354e1af4733b716b7b69f93244b3780f1b364c67c4b664c34ad0dcddff28e3 WatchSource:0}: Error finding container 37354e1af4733b716b7b69f93244b3780f1b364c67c4b664c34ad0dcddff28e3: Status 404 returned error can't find the container with id 37354e1af4733b716b7b69f93244b3780f1b364c67c4b664c34ad0dcddff28e3 Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.357420 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-658c75dcc7-mnxjh"] Feb 21 00:16:58 crc kubenswrapper[4730]: W0221 00:16:58.371215 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod158c1ae3_ea68_47d1_bc9e_654f1b566517.slice/crio-1ff3bcc049dd05c82426626991dadaf94e311e6d669c03c8ddd1714d4a040ad7 WatchSource:0}: Error finding container 1ff3bcc049dd05c82426626991dadaf94e311e6d669c03c8ddd1714d4a040ad7: Status 404 returned error can't find the container with id 1ff3bcc049dd05c82426626991dadaf94e311e6d669c03c8ddd1714d4a040ad7 Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.622581 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" event={"ID":"70369d12-f9ef-4028-9a45-62aad9cca5e5","Type":"ContainerStarted","Data":"37354e1af4733b716b7b69f93244b3780f1b364c67c4b664c34ad0dcddff28e3"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.624524 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" event={"ID":"7d174177-75da-4fb4-a195-a3e06804ba1e","Type":"ContainerStarted","Data":"08e4f760a0151d8d33abeb1409740414b83af6b38b6173d38a09d67395b85158"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.626068 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" event={"ID":"b9cd1a14-f993-46fc-816b-884442179c47","Type":"ContainerStarted","Data":"8902a4154610a319e2e89de2117fd9f018451e2e8bf3d39ece553fa88b12b8d8"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.626253 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.628115 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" event={"ID":"5e6832e1-eb7d-4f11-9e11-95c61cbf0934","Type":"ContainerStarted","Data":"b9001a27a891bb38c8f61f2a4196d1208b69014147a0d66674edb75b5cbc3d23"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.630252 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" event={"ID":"e7055108-18d7-4bc4-8644-a8d80b1605b3","Type":"ContainerStarted","Data":"90e4c98071f814c6747b81760a6c151cc01faaae79eef24d43b572f123bd3489"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.631107 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.632515 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" event={"ID":"158c1ae3-ea68-47d1-bc9e-654f1b566517","Type":"ContainerStarted","Data":"1ff3bcc049dd05c82426626991dadaf94e311e6d669c03c8ddd1714d4a040ad7"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.634068 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" event={"ID":"018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd","Type":"ContainerStarted","Data":"c93b1597c7030ad2164358e757dc6a6c025d44fb387eab543476a393e969d718"} Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.646883 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5" podStartSLOduration=2.183960312 podStartE2EDuration="13.646864232s" podCreationTimestamp="2026-02-21 00:16:45 +0000 UTC" firstStartedPulling="2026-02-21 00:16:46.567485828 +0000 UTC m=+658.858321558" lastFinishedPulling="2026-02-21 00:16:58.030389748 +0000 UTC m=+670.321225478" observedRunningTime="2026-02-21 00:16:58.638320255 +0000 UTC m=+670.929156015" watchObservedRunningTime="2026-02-21 00:16:58.646864232 +0000 UTC m=+670.937699962" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.662957 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.696820 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-kfkqw" podStartSLOduration=2.20157458 podStartE2EDuration="13.696806454s" podCreationTimestamp="2026-02-21 00:16:45 +0000 UTC" firstStartedPulling="2026-02-21 00:16:46.540330923 +0000 UTC m=+658.831166653" lastFinishedPulling="2026-02-21 00:16:58.035562797 +0000 UTC m=+670.326398527" observedRunningTime="2026-02-21 00:16:58.693445117 +0000 UTC m=+670.984280847" watchObservedRunningTime="2026-02-21 00:16:58.696806454 +0000 UTC m=+670.987642184" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.697720 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" podStartSLOduration=2.206258916 podStartE2EDuration="13.697715131s" podCreationTimestamp="2026-02-21 00:16:45 +0000 UTC" firstStartedPulling="2026-02-21 00:16:46.588713181 +0000 UTC m=+658.879548911" lastFinishedPulling="2026-02-21 00:16:58.080169396 +0000 UTC m=+670.371005126" observedRunningTime="2026-02-21 00:16:58.672977916 +0000 UTC m=+670.963813656" watchObservedRunningTime="2026-02-21 00:16:58.697715131 +0000 UTC m=+670.988550851" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.721049 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-wh7ls" podStartSLOduration=2.337402743 podStartE2EDuration="13.721029674s" podCreationTimestamp="2026-02-21 00:16:45 +0000 UTC" firstStartedPulling="2026-02-21 00:16:46.675488157 +0000 UTC m=+658.966323887" lastFinishedPulling="2026-02-21 00:16:58.059115088 +0000 UTC m=+670.349950818" observedRunningTime="2026-02-21 00:16:58.718894552 +0000 UTC m=+671.009730282" watchObservedRunningTime="2026-02-21 00:16:58.721029674 +0000 UTC m=+671.011865404" Feb 21 00:16:58 crc kubenswrapper[4730]: I0221 00:16:58.739915 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7" podStartSLOduration=2.061151714 podStartE2EDuration="13.739883008s" podCreationTimestamp="2026-02-21 00:16:45 +0000 UTC" firstStartedPulling="2026-02-21 00:16:46.358750849 +0000 UTC m=+658.649586579" lastFinishedPulling="2026-02-21 00:16:58.037482143 +0000 UTC m=+670.328317873" observedRunningTime="2026-02-21 00:16:58.737652474 +0000 UTC m=+671.028488204" watchObservedRunningTime="2026-02-21 00:16:58.739883008 +0000 UTC m=+671.030718738" Feb 21 00:17:01 crc kubenswrapper[4730]: I0221 00:17:01.656451 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" event={"ID":"158c1ae3-ea68-47d1-bc9e-654f1b566517","Type":"ContainerStarted","Data":"763439a1e5b663f7cd093d63515a62916befd059245e5857de4e3f54290d9247"} Feb 21 00:17:01 crc kubenswrapper[4730]: I0221 00:17:01.685559 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-658c75dcc7-mnxjh" podStartSLOduration=4.833977195 podStartE2EDuration="7.685539769s" podCreationTimestamp="2026-02-21 00:16:54 +0000 UTC" firstStartedPulling="2026-02-21 00:16:58.374239359 +0000 UTC m=+670.665075089" lastFinishedPulling="2026-02-21 00:17:01.225801933 +0000 UTC m=+673.516637663" observedRunningTime="2026-02-21 00:17:01.682670236 +0000 UTC m=+673.973505976" watchObservedRunningTime="2026-02-21 00:17:01.685539769 +0000 UTC m=+673.976375499" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.589895 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.591162 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.593343 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.593556 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.593846 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.594262 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.594520 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.594699 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-6hrr2" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.595677 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.595864 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.599790 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.606662 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719275 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719341 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719369 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719393 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719454 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719495 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719527 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719548 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719569 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719607 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719640 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719661 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719685 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9ad64692-bf81-4143-b748-f074c8b4e49a-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719708 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.719736 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821320 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821369 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821388 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821408 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9ad64692-bf81-4143-b748-f074c8b4e49a-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821425 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821445 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821469 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821499 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821517 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821538 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821557 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821584 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821599 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821613 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.821629 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.822218 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.822662 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.822917 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.823329 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.823415 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.823915 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.824163 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.827584 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/9ad64692-bf81-4143-b748-f074c8b4e49a-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.828260 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.829115 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.829118 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.831378 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.832565 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.832798 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/9ad64692-bf81-4143-b748-f074c8b4e49a-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.833158 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/9ad64692-bf81-4143-b748-f074c8b4e49a-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"9ad64692-bf81-4143-b748-f074c8b4e49a\") " pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:03 crc kubenswrapper[4730]: I0221 00:17:03.911843 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:06 crc kubenswrapper[4730]: I0221 00:17:06.240916 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-9vq9k" Feb 21 00:17:06 crc kubenswrapper[4730]: I0221 00:17:06.766860 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 21 00:17:07 crc kubenswrapper[4730]: I0221 00:17:07.737562 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" event={"ID":"70369d12-f9ef-4028-9a45-62aad9cca5e5","Type":"ContainerStarted","Data":"b266aed8491028df049325188aa5abe1fd544b006044fc5bb460eeb02c682899"} Feb 21 00:17:07 crc kubenswrapper[4730]: I0221 00:17:07.739292 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9ad64692-bf81-4143-b748-f074c8b4e49a","Type":"ContainerStarted","Data":"71ce70c4f5cc38452acbb406bb34e8222f1875ef5972db975cf16256c5c62ddd"} Feb 21 00:17:07 crc kubenswrapper[4730]: I0221 00:17:07.756713 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-7wv8x" podStartSLOduration=8.482790553 podStartE2EDuration="16.756695175s" podCreationTimestamp="2026-02-21 00:16:51 +0000 UTC" firstStartedPulling="2026-02-21 00:16:58.346843987 +0000 UTC m=+670.637679717" lastFinishedPulling="2026-02-21 00:17:06.620748599 +0000 UTC m=+678.911584339" observedRunningTime="2026-02-21 00:17:07.752078542 +0000 UTC m=+680.042914262" watchObservedRunningTime="2026-02-21 00:17:07.756695175 +0000 UTC m=+680.047530905" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.138828 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.140165 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.144311 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-sys-config" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.144355 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-global-ca" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.144912 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-dkmk9" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.163144 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-1-ca" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.190844 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.232445 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.232770 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.232913 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233062 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233161 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233271 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc6xf\" (UniqueName: \"kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233368 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233468 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233568 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233658 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233759 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.233828 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335099 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335148 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335164 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335212 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335257 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335277 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335307 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335322 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335345 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335366 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335385 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335402 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc6xf\" (UniqueName: \"kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335519 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.335715 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.336310 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.336377 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.336773 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.339550 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.341220 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.341681 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.346262 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.346381 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.357424 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc6xf\" (UniqueName: \"kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.364078 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-1-build\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:11 crc kubenswrapper[4730]: I0221 00:17:11.460964 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.061579 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:12 crc kubenswrapper[4730]: W0221 00:17:12.074024 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7c335b9e_69ea_4447_8b4f_729e483f4e05.slice/crio-24f0a30f7d3995ab03c0bab987e9822499ca629348127d36a4e8304bb63ec8b5 WatchSource:0}: Error finding container 24f0a30f7d3995ab03c0bab987e9822499ca629348127d36a4e8304bb63ec8b5: Status 404 returned error can't find the container with id 24f0a30f7d3995ab03c0bab987e9822499ca629348127d36a4e8304bb63ec8b5 Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.693470 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s"] Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.694723 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.696677 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.700808 4730 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-8r25z" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.701484 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.701975 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s"] Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.786782 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"7c335b9e-69ea-4447-8b4f-729e483f4e05","Type":"ContainerStarted","Data":"24f0a30f7d3995ab03c0bab987e9822499ca629348127d36a4e8304bb63ec8b5"} Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.852605 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c97ec86-76ba-4b4d-b267-1cee508b4394-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.852663 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7v96\" (UniqueName: \"kubernetes.io/projected/4c97ec86-76ba-4b4d-b267-1cee508b4394-kube-api-access-d7v96\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.954313 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c97ec86-76ba-4b4d-b267-1cee508b4394-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.954374 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7v96\" (UniqueName: \"kubernetes.io/projected/4c97ec86-76ba-4b4d-b267-1cee508b4394-kube-api-access-d7v96\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.954881 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/4c97ec86-76ba-4b4d-b267-1cee508b4394-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:12 crc kubenswrapper[4730]: I0221 00:17:12.972218 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7v96\" (UniqueName: \"kubernetes.io/projected/4c97ec86-76ba-4b4d-b267-1cee508b4394-kube-api-access-d7v96\") pod \"cert-manager-operator-controller-manager-5586865c96-w6k9s\" (UID: \"4c97ec86-76ba-4b4d-b267-1cee508b4394\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:13 crc kubenswrapper[4730]: I0221 00:17:13.010015 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" Feb 21 00:17:13 crc kubenswrapper[4730]: I0221 00:17:13.509684 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s"] Feb 21 00:17:13 crc kubenswrapper[4730]: W0221 00:17:13.517235 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4c97ec86_76ba_4b4d_b267_1cee508b4394.slice/crio-5a662eecbfeafa56324deb06681fd4ee2f28a8c3818a5eb7e36ad836373cb91a WatchSource:0}: Error finding container 5a662eecbfeafa56324deb06681fd4ee2f28a8c3818a5eb7e36ad836373cb91a: Status 404 returned error can't find the container with id 5a662eecbfeafa56324deb06681fd4ee2f28a8c3818a5eb7e36ad836373cb91a Feb 21 00:17:13 crc kubenswrapper[4730]: I0221 00:17:13.802256 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" event={"ID":"4c97ec86-76ba-4b4d-b267-1cee508b4394","Type":"ContainerStarted","Data":"5a662eecbfeafa56324deb06681fd4ee2f28a8c3818a5eb7e36ad836373cb91a"} Feb 21 00:17:22 crc kubenswrapper[4730]: I0221 00:17:22.325002 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.162103 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.167876 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.172891 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-ca" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.174674 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-global-ca" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.174994 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-2-sys-config" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.179683 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206540 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206599 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206621 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206637 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206654 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206673 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206722 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206882 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.206973 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.207014 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.207057 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.207092 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjfxd\" (UniqueName: \"kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307500 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307559 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307586 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307612 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307638 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307661 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307704 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308296 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308438 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308636 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308686 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.307732 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308882 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308885 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.308965 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309034 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309356 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309427 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309503 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309460 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjfxd\" (UniqueName: \"kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.309762 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.313613 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.320887 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.324924 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjfxd\" (UniqueName: \"kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd\") pod \"service-telemetry-operator-2-build\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:24 crc kubenswrapper[4730]: I0221 00:17:24.500977 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:26 crc kubenswrapper[4730]: E0221 00:17:26.049607 4730 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="registry.connect.redhat.com/elastic/elasticsearch:7.17.20" Feb 21 00:17:26 crc kubenswrapper[4730]: E0221 00:17:26.050541 4730 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:elastic-internal-init-filesystem,Image:registry.connect.redhat.com/elastic/elasticsearch:7.17.20,Command:[bash -c /mnt/elastic-internal/scripts/prepare-fs.sh],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:POD_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.name,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NODE_NAME,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:spec.nodeName,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:HEADLESS_SERVICE_NAME,Value:elasticsearch-es-default,ValueFrom:nil,},EnvVar{Name:PROBE_PASSWORD_PATH,Value:/mnt/elastic-internal/pod-mounted-users/elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:PROBE_USERNAME,Value:elastic-internal-probe,ValueFrom:nil,},EnvVar{Name:READINESS_PROBE_PROTOCOL,Value:https,ValueFrom:nil,},EnvVar{Name:NSS_SDB_USE_CACHE,Value:no,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Requests:ResourceList{cpu: {{100 -3} {} 100m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:downward-api,ReadOnly:true,MountPath:/mnt/elastic-internal/downward-api,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-bin-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-bin-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config,ReadOnly:true,MountPath:/mnt/elastic-internal/elasticsearch-config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-config-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-config-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-elasticsearch-plugins-local,ReadOnly:false,MountPath:/mnt/elastic-internal/elasticsearch-plugins-local,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-http-certificates,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/http-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-probe-user,ReadOnly:true,MountPath:/mnt/elastic-internal/pod-mounted-users,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-remote-certificate-authorities,ReadOnly:true,MountPath:/usr/share/elasticsearch/config/transport-remote-certs/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-scripts,ReadOnly:true,MountPath:/mnt/elastic-internal/scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-transport-certificates,ReadOnly:true,MountPath:/mnt/elastic-internal/transport-certificates,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-unicast-hosts,ReadOnly:true,MountPath:/mnt/elastic-internal/unicast-hosts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elastic-internal-xpack-file-realm,ReadOnly:true,MountPath:/mnt/elastic-internal/xpack-file-realm,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-data,ReadOnly:false,MountPath:/usr/share/elasticsearch/data,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:elasticsearch-logs,ReadOnly:false,MountPath:/usr/share/elasticsearch/logs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tmp-volume,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:*false,SELinuxOptions:nil,RunAsUser:*1000670000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod elasticsearch-es-default-0_service-telemetry(9ad64692-bf81-4143-b748-f074c8b4e49a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Feb 21 00:17:26 crc kubenswrapper[4730]: E0221 00:17:26.051792 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="9ad64692-bf81-4143-b748-f074c8b4e49a" Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.412836 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:26 crc kubenswrapper[4730]: W0221 00:17:26.418187 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9504ba0_908d_49cb_b523_69f363ec33b4.slice/crio-3bd5f2904f271fd60973cd2a9469a62239a1b1dc2ebb5f2687699759ffd8eb66 WatchSource:0}: Error finding container 3bd5f2904f271fd60973cd2a9469a62239a1b1dc2ebb5f2687699759ffd8eb66: Status 404 returned error can't find the container with id 3bd5f2904f271fd60973cd2a9469a62239a1b1dc2ebb5f2687699759ffd8eb66 Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.900557 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"f9504ba0-908d-49cb-b523-69f363ec33b4","Type":"ContainerStarted","Data":"7e85e34ed2bacfe6bab18eef74489a7c4c84a099bb42a36702b1a6862d7821e7"} Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.901120 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"f9504ba0-908d-49cb-b523-69f363ec33b4","Type":"ContainerStarted","Data":"3bd5f2904f271fd60973cd2a9469a62239a1b1dc2ebb5f2687699759ffd8eb66"} Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.903241 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" event={"ID":"4c97ec86-76ba-4b4d-b267-1cee508b4394","Type":"ContainerStarted","Data":"ebfcab69f01af0780aed1a5d160db449dcfcecab8c00271df148e1bbc5569cd9"} Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.906193 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"7c335b9e-69ea-4447-8b4f-729e483f4e05","Type":"ContainerStarted","Data":"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b"} Feb 21 00:17:26 crc kubenswrapper[4730]: I0221 00:17:26.906308 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-1-build" podUID="7c335b9e-69ea-4447-8b4f-729e483f4e05" containerName="manage-dockerfile" containerID="cri-o://6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b" gracePeriod=30 Feb 21 00:17:26 crc kubenswrapper[4730]: E0221 00:17:26.923378 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="9ad64692-bf81-4143-b748-f074c8b4e49a" Feb 21 00:17:27 crc kubenswrapper[4730]: E0221 00:17:27.000410 4730 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=682249096903815999, SKID=, AKID=D2:AC:BB:E1:C4:4E:57:A5:30:EF:CB:DD:40:C4:E2:40:B5:C6:6A:67 failed: x509: certificate signed by unknown authority" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.051048 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-w6k9s" podStartSLOduration=2.708840332 podStartE2EDuration="15.051033488s" podCreationTimestamp="2026-02-21 00:17:12 +0000 UTC" firstStartedPulling="2026-02-21 00:17:13.520882895 +0000 UTC m=+685.811718625" lastFinishedPulling="2026-02-21 00:17:25.863076051 +0000 UTC m=+698.153911781" observedRunningTime="2026-02-21 00:17:27.014330648 +0000 UTC m=+699.305166378" watchObservedRunningTime="2026-02-21 00:17:27.051033488 +0000 UTC m=+699.341869218" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.202514 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.240128 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.291485 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_7c335b9e-69ea-4447-8b4f-729e483f4e05/manage-dockerfile/0.log" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.291989 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348131 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348182 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348235 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc6xf\" (UniqueName: \"kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348273 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348302 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348326 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348380 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348422 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348452 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348474 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348497 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348550 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root\") pod \"7c335b9e-69ea-4447-8b4f-729e483f4e05\" (UID: \"7c335b9e-69ea-4447-8b4f-729e483f4e05\") " Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348403 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348658 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348769 4730 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348784 4730 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/7c335b9e-69ea-4447-8b4f-729e483f4e05-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348974 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.348992 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.349521 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.349547 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.349662 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.349724 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.349955 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.353387 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-push") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "builder-dockercfg-dkmk9-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.353641 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf" (OuterVolumeSpecName: "kube-api-access-bc6xf") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "kube-api-access-bc6xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.356451 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-pull") pod "7c335b9e-69ea-4447-8b4f-729e483f4e05" (UID: "7c335b9e-69ea-4447-8b4f-729e483f4e05"). InnerVolumeSpecName "builder-dockercfg-dkmk9-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449463 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449504 4730 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449512 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc6xf\" (UniqueName: \"kubernetes.io/projected/7c335b9e-69ea-4447-8b4f-729e483f4e05-kube-api-access-bc6xf\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449522 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-push\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449531 4730 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449540 4730 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449548 4730 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449555 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/7c335b9e-69ea-4447-8b4f-729e483f4e05-builder-dockercfg-dkmk9-pull\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449563 4730 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7c335b9e-69ea-4447-8b4f-729e483f4e05-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.449572 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/7c335b9e-69ea-4447-8b4f-729e483f4e05-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913416 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-1-build_7c335b9e-69ea-4447-8b4f-729e483f4e05/manage-dockerfile/0.log" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913476 4730 generic.go:334] "Generic (PLEG): container finished" podID="7c335b9e-69ea-4447-8b4f-729e483f4e05" containerID="6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b" exitCode=1 Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913553 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-1-build" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913561 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"7c335b9e-69ea-4447-8b4f-729e483f4e05","Type":"ContainerDied","Data":"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b"} Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913601 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-1-build" event={"ID":"7c335b9e-69ea-4447-8b4f-729e483f4e05","Type":"ContainerDied","Data":"24f0a30f7d3995ab03c0bab987e9822499ca629348127d36a4e8304bb63ec8b5"} Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.913620 4730 scope.go:117] "RemoveContainer" containerID="6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b" Feb 21 00:17:27 crc kubenswrapper[4730]: E0221 00:17:27.915227 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="9ad64692-bf81-4143-b748-f074c8b4e49a" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.942815 4730 scope.go:117] "RemoveContainer" containerID="6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b" Feb 21 00:17:27 crc kubenswrapper[4730]: E0221 00:17:27.943214 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b\": container with ID starting with 6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b not found: ID does not exist" containerID="6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.943275 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b"} err="failed to get container status \"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b\": rpc error: code = NotFound desc = could not find container \"6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b\": container with ID starting with 6b5fbab867fba90ca1937cc2398b61a18d37894427f03e8875df4c49cebeb02b not found: ID does not exist" Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.979367 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:27 crc kubenswrapper[4730]: I0221 00:17:27.991179 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-1-build"] Feb 21 00:17:28 crc kubenswrapper[4730]: I0221 00:17:28.031513 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:28 crc kubenswrapper[4730]: I0221 00:17:28.441417 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c335b9e-69ea-4447-8b4f-729e483f4e05" path="/var/lib/kubelet/pods/7c335b9e-69ea-4447-8b4f-729e483f4e05/volumes" Feb 21 00:17:28 crc kubenswrapper[4730]: I0221 00:17:28.920419 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-2-build" podUID="f9504ba0-908d-49cb-b523-69f363ec33b4" containerName="git-clone" containerID="cri-o://7e85e34ed2bacfe6bab18eef74489a7c4c84a099bb42a36702b1a6862d7821e7" gracePeriod=30 Feb 21 00:17:28 crc kubenswrapper[4730]: E0221 00:17:28.924827 4730 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"elastic-internal-init-filesystem\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.connect.redhat.com/elastic/elasticsearch:7.17.20\\\"\"" pod="service-telemetry/elasticsearch-es-default-0" podUID="9ad64692-bf81-4143-b748-f074c8b4e49a" Feb 21 00:17:29 crc kubenswrapper[4730]: I0221 00:17:29.926641 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_f9504ba0-908d-49cb-b523-69f363ec33b4/git-clone/0.log" Feb 21 00:17:29 crc kubenswrapper[4730]: I0221 00:17:29.926687 4730 generic.go:334] "Generic (PLEG): container finished" podID="f9504ba0-908d-49cb-b523-69f363ec33b4" containerID="7e85e34ed2bacfe6bab18eef74489a7c4c84a099bb42a36702b1a6862d7821e7" exitCode=1 Feb 21 00:17:29 crc kubenswrapper[4730]: I0221 00:17:29.926713 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"f9504ba0-908d-49cb-b523-69f363ec33b4","Type":"ContainerDied","Data":"7e85e34ed2bacfe6bab18eef74489a7c4c84a099bb42a36702b1a6862d7821e7"} Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.111253 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-mtvcd"] Feb 21 00:17:33 crc kubenswrapper[4730]: E0221 00:17:33.112368 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c335b9e-69ea-4447-8b4f-729e483f4e05" containerName="manage-dockerfile" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.112431 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c335b9e-69ea-4447-8b4f-729e483f4e05" containerName="manage-dockerfile" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.112579 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c335b9e-69ea-4447-8b4f-729e483f4e05" containerName="manage-dockerfile" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.113014 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.115133 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.115587 4730 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-fchz6" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.115752 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.122118 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-mtvcd"] Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.186880 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gg4jz\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-kube-api-access-gg4jz\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.187092 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.288542 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.288942 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gg4jz\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-kube-api-access-gg4jz\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.308209 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gg4jz\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-kube-api-access-gg4jz\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.308443 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/359d616a-dce4-4e42-bec5-b4fd29472159-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-mtvcd\" (UID: \"359d616a-dce4-4e42-bec5-b4fd29472159\") " pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.432411 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.522337 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_f9504ba0-908d-49cb-b523-69f363ec33b4/git-clone/0.log" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.522464 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.557557 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-4vngm"] Feb 21 00:17:33 crc kubenswrapper[4730]: E0221 00:17:33.558147 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9504ba0-908d-49cb-b523-69f363ec33b4" containerName="git-clone" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.558167 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9504ba0-908d-49cb-b523-69f363ec33b4" containerName="git-clone" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.558322 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9504ba0-908d-49cb-b523-69f363ec33b4" containerName="git-clone" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.558778 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.563372 4730 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tsvtz" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.574773 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-4vngm"] Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592189 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592263 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592302 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592316 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592350 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592369 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592400 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592420 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592452 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592470 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592489 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjfxd\" (UniqueName: \"kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592509 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root\") pod \"f9504ba0-908d-49cb-b523-69f363ec33b4\" (UID: \"f9504ba0-908d-49cb-b523-69f363ec33b4\") " Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592666 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.592865 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.593086 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.593370 4730 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.593397 4730 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f9504ba0-908d-49cb-b523-69f363ec33b4-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.593406 4730 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594068 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594080 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594085 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594260 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594475 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.594931 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.595234 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-pull") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "builder-dockercfg-dkmk9-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.595273 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-push") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "builder-dockercfg-dkmk9-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.598167 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd" (OuterVolumeSpecName: "kube-api-access-hjfxd") pod "f9504ba0-908d-49cb-b523-69f363ec33b4" (UID: "f9504ba0-908d-49cb-b523-69f363ec33b4"). InnerVolumeSpecName "kube-api-access-hjfxd". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.694941 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695149 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g96br\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-kube-api-access-g96br\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695315 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-pull\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695327 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695337 4730 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695348 4730 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695358 4730 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695368 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjfxd\" (UniqueName: \"kubernetes.io/projected/f9504ba0-908d-49cb-b523-69f363ec33b4-kube-api-access-hjfxd\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695378 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f9504ba0-908d-49cb-b523-69f363ec33b4-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695386 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f9504ba0-908d-49cb-b523-69f363ec33b4-builder-dockercfg-dkmk9-push\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.695394 4730 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f9504ba0-908d-49cb-b523-69f363ec33b4-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.796642 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g96br\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-kube-api-access-g96br\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.796764 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.821338 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.827455 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g96br\" (UniqueName: \"kubernetes.io/projected/1a9c9953-4d26-4895-bd21-d7e9c7f95a71-kube-api-access-g96br\") pod \"cert-manager-cainjector-5545bd876-4vngm\" (UID: \"1a9c9953-4d26-4895-bd21-d7e9c7f95a71\") " pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.869469 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-mtvcd"] Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.885643 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.950988 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-2-build_f9504ba0-908d-49cb-b523-69f363ec33b4/git-clone/0.log" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.951120 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-2-build" event={"ID":"f9504ba0-908d-49cb-b523-69f363ec33b4","Type":"ContainerDied","Data":"3bd5f2904f271fd60973cd2a9469a62239a1b1dc2ebb5f2687699759ffd8eb66"} Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.951145 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-2-build" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.951170 4730 scope.go:117] "RemoveContainer" containerID="7e85e34ed2bacfe6bab18eef74489a7c4c84a099bb42a36702b1a6862d7821e7" Feb 21 00:17:33 crc kubenswrapper[4730]: I0221 00:17:33.958094 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" event={"ID":"359d616a-dce4-4e42-bec5-b4fd29472159","Type":"ContainerStarted","Data":"cc7c5a010be3f23e9b69165eafddd816860079ead91276bd397418d884c0a0be"} Feb 21 00:17:34 crc kubenswrapper[4730]: I0221 00:17:34.013626 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:34 crc kubenswrapper[4730]: I0221 00:17:34.032926 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-2-build"] Feb 21 00:17:34 crc kubenswrapper[4730]: I0221 00:17:34.374085 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-4vngm"] Feb 21 00:17:34 crc kubenswrapper[4730]: W0221 00:17:34.379155 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a9c9953_4d26_4895_bd21_d7e9c7f95a71.slice/crio-2e3ada60ea61b4762477bc4463e907ebedf8c0cd48bcbe33be61f30155ee67ad WatchSource:0}: Error finding container 2e3ada60ea61b4762477bc4463e907ebedf8c0cd48bcbe33be61f30155ee67ad: Status 404 returned error can't find the container with id 2e3ada60ea61b4762477bc4463e907ebedf8c0cd48bcbe33be61f30155ee67ad Feb 21 00:17:34 crc kubenswrapper[4730]: I0221 00:17:34.444372 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9504ba0-908d-49cb-b523-69f363ec33b4" path="/var/lib/kubelet/pods/f9504ba0-908d-49cb-b523-69f363ec33b4/volumes" Feb 21 00:17:34 crc kubenswrapper[4730]: I0221 00:17:34.966013 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" event={"ID":"1a9c9953-4d26-4895-bd21-d7e9c7f95a71","Type":"ContainerStarted","Data":"2e3ada60ea61b4762477bc4463e907ebedf8c0cd48bcbe33be61f30155ee67ad"} Feb 21 00:17:38 crc kubenswrapper[4730]: I0221 00:17:38.996374 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" event={"ID":"359d616a-dce4-4e42-bec5-b4fd29472159","Type":"ContainerStarted","Data":"e08f34e43f642ed2e803370452d4433c7d5b5f8275fd31d76801e1c22a187cba"} Feb 21 00:17:38 crc kubenswrapper[4730]: I0221 00:17:38.996929 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:38 crc kubenswrapper[4730]: I0221 00:17:38.999402 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" event={"ID":"1a9c9953-4d26-4895-bd21-d7e9c7f95a71","Type":"ContainerStarted","Data":"c1018fe96efa41a45cfb87f478fb4498ad4b4414cfc8b65084a1c41e8e80ff18"} Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.014702 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" podStartSLOduration=1.461471743 podStartE2EDuration="6.014685161s" podCreationTimestamp="2026-02-21 00:17:33 +0000 UTC" firstStartedPulling="2026-02-21 00:17:33.88209274 +0000 UTC m=+706.172928490" lastFinishedPulling="2026-02-21 00:17:38.435306178 +0000 UTC m=+710.726141908" observedRunningTime="2026-02-21 00:17:39.01257857 +0000 UTC m=+711.303414320" watchObservedRunningTime="2026-02-21 00:17:39.014685161 +0000 UTC m=+711.305520901" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.041147 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-4vngm" podStartSLOduration=2.002217301 podStartE2EDuration="6.041125245s" podCreationTimestamp="2026-02-21 00:17:33 +0000 UTC" firstStartedPulling="2026-02-21 00:17:34.380564867 +0000 UTC m=+706.671400597" lastFinishedPulling="2026-02-21 00:17:38.419472801 +0000 UTC m=+710.710308541" observedRunningTime="2026-02-21 00:17:39.028977804 +0000 UTC m=+711.319813534" watchObservedRunningTime="2026-02-21 00:17:39.041125245 +0000 UTC m=+711.331960975" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.455467 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.456810 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.458540 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-sys-config" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.465291 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-dkmk9" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.465298 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-global-ca" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.466162 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-3-ca" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.483584 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.579601 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.579885 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580099 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580220 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580390 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580476 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580496 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580533 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580565 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580603 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk7w2\" (UniqueName: \"kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580623 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.580643 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.610851 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-scgp8"] Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.611538 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.613322 4730 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-t875z" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.625143 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-scgp8"] Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681675 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681726 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681747 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk7w2\" (UniqueName: \"kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681766 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681785 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681804 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-bound-sa-token\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681823 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681841 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681873 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndm67\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-kube-api-access-ndm67\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681895 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681955 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.681982 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682016 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682031 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682334 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682442 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682765 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.682807 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.683689 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.686152 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.686563 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.686641 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.686748 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.687428 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.689125 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.701813 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk7w2\" (UniqueName: \"kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2\") pod \"service-telemetry-operator-3-build\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.776165 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.783288 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndm67\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-kube-api-access-ndm67\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.783450 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-bound-sa-token\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.802891 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-bound-sa-token\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.808510 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndm67\" (UniqueName: \"kubernetes.io/projected/3d8693e2-17ec-40dd-b2e1-ccc39aa71703-kube-api-access-ndm67\") pod \"cert-manager-545d4d4674-scgp8\" (UID: \"3d8693e2-17ec-40dd-b2e1-ccc39aa71703\") " pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:39 crc kubenswrapper[4730]: I0221 00:17:39.938163 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-scgp8" Feb 21 00:17:40 crc kubenswrapper[4730]: W0221 00:17:40.221257 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5905c38_a79e_4c66_bb7b_cbdba324bb16.slice/crio-1306ca407aeeb0435a37292bd32eba72308bce466008deed8cf46d63f1991189 WatchSource:0}: Error finding container 1306ca407aeeb0435a37292bd32eba72308bce466008deed8cf46d63f1991189: Status 404 returned error can't find the container with id 1306ca407aeeb0435a37292bd32eba72308bce466008deed8cf46d63f1991189 Feb 21 00:17:40 crc kubenswrapper[4730]: I0221 00:17:40.222338 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:40 crc kubenswrapper[4730]: I0221 00:17:40.357317 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-scgp8"] Feb 21 00:17:40 crc kubenswrapper[4730]: W0221 00:17:40.366732 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d8693e2_17ec_40dd_b2e1_ccc39aa71703.slice/crio-9acbe795762622ae6c53003a3e9e5b3d923f29301671a61da72ca441755a8bb3 WatchSource:0}: Error finding container 9acbe795762622ae6c53003a3e9e5b3d923f29301671a61da72ca441755a8bb3: Status 404 returned error can't find the container with id 9acbe795762622ae6c53003a3e9e5b3d923f29301671a61da72ca441755a8bb3 Feb 21 00:17:41 crc kubenswrapper[4730]: I0221 00:17:41.013801 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"f5905c38-a79e-4c66-bb7b-cbdba324bb16","Type":"ContainerStarted","Data":"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89"} Feb 21 00:17:41 crc kubenswrapper[4730]: I0221 00:17:41.014713 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"f5905c38-a79e-4c66-bb7b-cbdba324bb16","Type":"ContainerStarted","Data":"1306ca407aeeb0435a37292bd32eba72308bce466008deed8cf46d63f1991189"} Feb 21 00:17:41 crc kubenswrapper[4730]: I0221 00:17:41.015831 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-scgp8" event={"ID":"3d8693e2-17ec-40dd-b2e1-ccc39aa71703","Type":"ContainerStarted","Data":"1d2e0cd777daafaa6b6095c028ea4a2fa92a3ce149e9f68efd86b0ff15a131b3"} Feb 21 00:17:41 crc kubenswrapper[4730]: I0221 00:17:41.015884 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-scgp8" event={"ID":"3d8693e2-17ec-40dd-b2e1-ccc39aa71703","Type":"ContainerStarted","Data":"9acbe795762622ae6c53003a3e9e5b3d923f29301671a61da72ca441755a8bb3"} Feb 21 00:17:41 crc kubenswrapper[4730]: I0221 00:17:41.079050 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-scgp8" podStartSLOduration=2.079009098 podStartE2EDuration="2.079009098s" podCreationTimestamp="2026-02-21 00:17:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-02-21 00:17:41.07385654 +0000 UTC m=+713.364692300" watchObservedRunningTime="2026-02-21 00:17:41.079009098 +0000 UTC m=+713.369844848" Feb 21 00:17:41 crc kubenswrapper[4730]: E0221 00:17:41.082604 4730 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=682249096903815999, SKID=, AKID=D2:AC:BB:E1:C4:4E:57:A5:30:EF:CB:DD:40:C4:E2:40:B5:C6:6A:67 failed: x509: certificate signed by unknown authority" Feb 21 00:17:42 crc kubenswrapper[4730]: I0221 00:17:42.126584 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:43 crc kubenswrapper[4730]: I0221 00:17:43.030631 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9ad64692-bf81-4143-b748-f074c8b4e49a","Type":"ContainerStarted","Data":"62aa9557bccf9f7f906ce8630243034567663dc1fdd4b2e008d9af86da25c79f"} Feb 21 00:17:43 crc kubenswrapper[4730]: I0221 00:17:43.435629 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-mtvcd" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.037770 4730 generic.go:334] "Generic (PLEG): container finished" podID="9ad64692-bf81-4143-b748-f074c8b4e49a" containerID="62aa9557bccf9f7f906ce8630243034567663dc1fdd4b2e008d9af86da25c79f" exitCode=0 Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.038008 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-3-build" podUID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" containerName="git-clone" containerID="cri-o://c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89" gracePeriod=30 Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.038081 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9ad64692-bf81-4143-b748-f074c8b4e49a","Type":"ContainerDied","Data":"62aa9557bccf9f7f906ce8630243034567663dc1fdd4b2e008d9af86da25c79f"} Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.415827 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_f5905c38-a79e-4c66-bb7b-cbdba324bb16/git-clone/0.log" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.416201 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565247 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565316 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565361 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565392 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565410 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565446 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565474 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565498 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565528 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565562 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565587 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk7w2\" (UniqueName: \"kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565609 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets\") pod \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\" (UID: \"f5905c38-a79e-4c66-bb7b-cbdba324bb16\") " Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565664 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565843 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565872 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.565962 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566058 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566070 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566216 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566245 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566271 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.566421 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.571177 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-pull") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "builder-dockercfg-dkmk9-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.571218 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2" (OuterVolumeSpecName: "kube-api-access-kk7w2") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "kube-api-access-kk7w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.571627 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-push") pod "f5905c38-a79e-4c66-bb7b-cbdba324bb16" (UID: "f5905c38-a79e-4c66-bb7b-cbdba324bb16"). InnerVolumeSpecName "builder-dockercfg-dkmk9-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.666951 4730 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667030 4730 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667057 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk7w2\" (UniqueName: \"kubernetes.io/projected/f5905c38-a79e-4c66-bb7b-cbdba324bb16-kube-api-access-kk7w2\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667080 4730 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/f5905c38-a79e-4c66-bb7b-cbdba324bb16-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667102 4730 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667126 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-push\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667166 4730 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667191 4730 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667213 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/f5905c38-a79e-4c66-bb7b-cbdba324bb16-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667236 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/f5905c38-a79e-4c66-bb7b-cbdba324bb16-builder-dockercfg-dkmk9-pull\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:44 crc kubenswrapper[4730]: I0221 00:17:44.667258 4730 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/f5905c38-a79e-4c66-bb7b-cbdba324bb16-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.044785 4730 generic.go:334] "Generic (PLEG): container finished" podID="9ad64692-bf81-4143-b748-f074c8b4e49a" containerID="da4eb0e9a25815e65e58d7be20310b3ef45fb602f7304058d61ecb07e6f6876b" exitCode=0 Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.044862 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9ad64692-bf81-4143-b748-f074c8b4e49a","Type":"ContainerDied","Data":"da4eb0e9a25815e65e58d7be20310b3ef45fb602f7304058d61ecb07e6f6876b"} Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046456 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-3-build_f5905c38-a79e-4c66-bb7b-cbdba324bb16/git-clone/0.log" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046496 4730 generic.go:334] "Generic (PLEG): container finished" podID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" containerID="c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89" exitCode=1 Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046518 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"f5905c38-a79e-4c66-bb7b-cbdba324bb16","Type":"ContainerDied","Data":"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89"} Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046539 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-3-build" event={"ID":"f5905c38-a79e-4c66-bb7b-cbdba324bb16","Type":"ContainerDied","Data":"1306ca407aeeb0435a37292bd32eba72308bce466008deed8cf46d63f1991189"} Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046557 4730 scope.go:117] "RemoveContainer" containerID="c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.046579 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-3-build" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.072355 4730 scope.go:117] "RemoveContainer" containerID="c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89" Feb 21 00:17:45 crc kubenswrapper[4730]: E0221 00:17:45.072835 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89\": container with ID starting with c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89 not found: ID does not exist" containerID="c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.072917 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89"} err="failed to get container status \"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89\": rpc error: code = NotFound desc = could not find container \"c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89\": container with ID starting with c5a04420cd58354e9737b532afe752f2bd1d1833016f516bd8f27c1849a2da89 not found: ID does not exist" Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.090851 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:45 crc kubenswrapper[4730]: I0221 00:17:45.102766 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-3-build"] Feb 21 00:17:46 crc kubenswrapper[4730]: I0221 00:17:46.063135 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"9ad64692-bf81-4143-b748-f074c8b4e49a","Type":"ContainerStarted","Data":"49dd1157966aa882a220cce811d44d9ad7023fc4bc8523b84b893e79786309c8"} Feb 21 00:17:46 crc kubenswrapper[4730]: I0221 00:17:46.064976 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:17:46 crc kubenswrapper[4730]: I0221 00:17:46.447811 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" path="/var/lib/kubelet/pods/f5905c38-a79e-4c66-bb7b-cbdba324bb16/volumes" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.645709 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=15.347399538 podStartE2EDuration="50.645687175s" podCreationTimestamp="2026-02-21 00:17:03 +0000 UTC" firstStartedPulling="2026-02-21 00:17:06.783852179 +0000 UTC m=+679.074687899" lastFinishedPulling="2026-02-21 00:17:42.082139806 +0000 UTC m=+714.372975536" observedRunningTime="2026-02-21 00:17:46.101649116 +0000 UTC m=+718.392484846" watchObservedRunningTime="2026-02-21 00:17:53.645687175 +0000 UTC m=+725.936522925" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.650123 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:53 crc kubenswrapper[4730]: E0221 00:17:53.650412 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" containerName="git-clone" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.650427 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" containerName="git-clone" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.650591 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5905c38-a79e-4c66-bb7b-cbdba324bb16" containerName="git-clone" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.651625 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.654558 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-sys-config" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.655164 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-global-ca" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.655160 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-dkmk9" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.657592 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-4-ca" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.682022 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785667 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785748 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785778 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785799 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785836 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785949 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.785990 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.786014 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.786037 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.786079 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hv78j\" (UniqueName: \"kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.786101 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.786120 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.887721 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.887788 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.887828 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.887863 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.887957 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.888065 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.888120 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.888792 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.888890 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.889351 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.889398 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.888698 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.889522 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.890033 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.890803 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.890991 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hv78j\" (UniqueName: \"kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.891031 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.891072 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.891165 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.891801 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.892099 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.897371 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.903140 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.912387 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hv78j\" (UniqueName: \"kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j\") pod \"service-telemetry-operator-4-build\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:53 crc kubenswrapper[4730]: I0221 00:17:53.974582 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:54 crc kubenswrapper[4730]: I0221 00:17:54.262634 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:54 crc kubenswrapper[4730]: W0221 00:17:54.278099 4730 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c5940c0_aab6_4b08_bbdd_f453d6c51b68.slice/crio-0bdaf426b0770895855bb9a797647b853c6e94a819fcd2608bfd9d680b8e13a9 WatchSource:0}: Error finding container 0bdaf426b0770895855bb9a797647b853c6e94a819fcd2608bfd9d680b8e13a9: Status 404 returned error can't find the container with id 0bdaf426b0770895855bb9a797647b853c6e94a819fcd2608bfd9d680b8e13a9 Feb 21 00:17:55 crc kubenswrapper[4730]: I0221 00:17:55.133691 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"6c5940c0-aab6-4b08-bbdd-f453d6c51b68","Type":"ContainerStarted","Data":"0bdaf426b0770895855bb9a797647b853c6e94a819fcd2608bfd9d680b8e13a9"} Feb 21 00:17:56 crc kubenswrapper[4730]: I0221 00:17:56.146134 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"6c5940c0-aab6-4b08-bbdd-f453d6c51b68","Type":"ContainerStarted","Data":"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532"} Feb 21 00:17:56 crc kubenswrapper[4730]: E0221 00:17:56.212236 4730 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=682249096903815999, SKID=, AKID=D2:AC:BB:E1:C4:4E:57:A5:30:EF:CB:DD:40:C4:E2:40:B5:C6:6A:67 failed: x509: certificate signed by unknown authority" Feb 21 00:17:57 crc kubenswrapper[4730]: I0221 00:17:57.248674 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.160100 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-4-build" podUID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" containerName="git-clone" containerID="cri-o://8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532" gracePeriod=30 Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.587507 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_6c5940c0-aab6-4b08-bbdd-f453d6c51b68/git-clone/0.log" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.587789 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659711 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659754 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659781 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659832 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hv78j\" (UniqueName: \"kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659858 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659891 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659974 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660002 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660024 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.659895 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660012 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660368 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660591 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660083 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660669 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660691 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660751 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles\") pod \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\" (UID: \"6c5940c0-aab6-4b08-bbdd-f453d6c51b68\") " Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660844 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.660984 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661167 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661325 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661428 4730 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661454 4730 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661472 4730 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661491 4730 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661507 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661524 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661543 4730 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661561 4730 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.661577 4730 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.664852 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j" (OuterVolumeSpecName: "kube-api-access-hv78j") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "kube-api-access-hv78j". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.676217 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-push") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "builder-dockercfg-dkmk9-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.677144 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-pull") pod "6c5940c0-aab6-4b08-bbdd-f453d6c51b68" (UID: "6c5940c0-aab6-4b08-bbdd-f453d6c51b68"). InnerVolumeSpecName "builder-dockercfg-dkmk9-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.763003 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-pull\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.763060 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hv78j\" (UniqueName: \"kubernetes.io/projected/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-kube-api-access-hv78j\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:58 crc kubenswrapper[4730]: I0221 00:17:58.763082 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/6c5940c0-aab6-4b08-bbdd-f453d6c51b68-builder-dockercfg-dkmk9-push\") on node \"crc\" DevicePath \"\"" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.170731 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-4-build_6c5940c0-aab6-4b08-bbdd-f453d6c51b68/git-clone/0.log" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.170777 4730 generic.go:334] "Generic (PLEG): container finished" podID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" containerID="8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532" exitCode=1 Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.170809 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"6c5940c0-aab6-4b08-bbdd-f453d6c51b68","Type":"ContainerDied","Data":"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532"} Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.170838 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-4-build" event={"ID":"6c5940c0-aab6-4b08-bbdd-f453d6c51b68","Type":"ContainerDied","Data":"0bdaf426b0770895855bb9a797647b853c6e94a819fcd2608bfd9d680b8e13a9"} Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.170860 4730 scope.go:117] "RemoveContainer" containerID="8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.171011 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-4-build" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.202196 4730 scope.go:117] "RemoveContainer" containerID="8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532" Feb 21 00:17:59 crc kubenswrapper[4730]: E0221 00:17:59.202581 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532\": container with ID starting with 8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532 not found: ID does not exist" containerID="8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.202624 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532"} err="failed to get container status \"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532\": rpc error: code = NotFound desc = could not find container \"8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532\": container with ID starting with 8ea5c2f3fbd5411bda832306a65bbd2db5bb76b5c03623a828ea63584959d532 not found: ID does not exist" Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.228981 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.232640 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-4-build"] Feb 21 00:17:59 crc kubenswrapper[4730]: I0221 00:17:59.621143 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Feb 21 00:18:00 crc kubenswrapper[4730]: I0221 00:18:00.446426 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" path="/var/lib/kubelet/pods/6c5940c0-aab6-4b08-bbdd-f453d6c51b68/volumes" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.716244 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:08 crc kubenswrapper[4730]: E0221 00:18:08.717237 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" containerName="git-clone" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.717258 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" containerName="git-clone" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.717441 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c5940c0-aab6-4b08-bbdd-f453d6c51b68" containerName="git-clone" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.718922 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.723031 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-ca" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.723429 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-sys-config" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.723634 4730 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"builder-dockercfg-dkmk9" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.724043 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"service-telemetry-operator-5-global-ca" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.738691 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823305 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823342 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823364 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823382 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823398 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823412 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzjhv\" (UniqueName: \"kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823432 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823527 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823598 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823644 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823679 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.823814 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.925984 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzjhv\" (UniqueName: \"kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926051 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926084 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926119 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926149 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926174 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926198 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926237 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926261 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926299 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926320 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926341 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926711 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926795 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.926864 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.927121 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.927459 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.928004 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.928100 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.928276 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.928939 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.938076 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.947703 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:08 crc kubenswrapper[4730]: I0221 00:18:08.948295 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzjhv\" (UniqueName: \"kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv\") pod \"service-telemetry-operator-5-build\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:09 crc kubenswrapper[4730]: I0221 00:18:09.040529 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:09 crc kubenswrapper[4730]: I0221 00:18:09.300159 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:10 crc kubenswrapper[4730]: I0221 00:18:10.264877 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a","Type":"ContainerStarted","Data":"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c"} Feb 21 00:18:10 crc kubenswrapper[4730]: I0221 00:18:10.265314 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a","Type":"ContainerStarted","Data":"a2a88ed3fc40a76129c8f1530731ac8404ca788dfb4c2831b319858e74e78610"} Feb 21 00:18:10 crc kubenswrapper[4730]: E0221 00:18:10.344742 4730 server.go:309] "Unable to authenticate the request due to an error" err="verifying certificate SN=682249096903815999, SKID=, AKID=D2:AC:BB:E1:C4:4E:57:A5:30:EF:CB:DD:40:C4:E2:40:B5:C6:6A:67 failed: x509: certificate signed by unknown authority" Feb 21 00:18:11 crc kubenswrapper[4730]: I0221 00:18:11.397569 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.281390 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="service-telemetry/service-telemetry-operator-5-build" podUID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" containerName="git-clone" containerID="cri-o://79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c" gracePeriod=30 Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.643598 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_8ede235c-a0a4-41df-9ed1-d8bbcbd8472a/git-clone/0.log" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.644060 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.809385 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.809786 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache" (OuterVolumeSpecName: "build-blob-cache") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "build-blob-cache". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.809854 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.809941 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.809995 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810038 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir" (OuterVolumeSpecName: "buildcachedir") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "buildcachedir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810069 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzjhv\" (UniqueName: \"kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810146 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810273 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810347 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810409 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810490 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810802 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles" (OuterVolumeSpecName: "build-ca-bundles") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "build-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810848 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.810959 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets\") pod \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\" (UID: \"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a\") " Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811272 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run" (OuterVolumeSpecName: "container-storage-run") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "container-storage-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811387 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets" (OuterVolumeSpecName: "node-pullsecrets") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "node-pullsecrets". PluginName "kubernetes.io/host-path", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811412 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir" (OuterVolumeSpecName: "buildworkdir") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "buildworkdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811417 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs" (OuterVolumeSpecName: "build-system-configs") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "build-system-configs". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811602 4730 reconciler_common.go:293] "Volume detached for volume \"buildworkdir\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildworkdir\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811622 4730 reconciler_common.go:293] "Volume detached for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-node-pullsecrets\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811633 4730 reconciler_common.go:293] "Volume detached for volume \"build-blob-cache\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-blob-cache\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811642 4730 reconciler_common.go:293] "Volume detached for volume \"buildcachedir\" (UniqueName: \"kubernetes.io/host-path/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-buildcachedir\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811651 4730 reconciler_common.go:293] "Volume detached for volume \"build-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811660 4730 reconciler_common.go:293] "Volume detached for volume \"build-system-configs\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-system-configs\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811669 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-run\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-run\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811663 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles" (OuterVolumeSpecName: "build-proxy-ca-bundles") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "build-proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.811806 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root" (OuterVolumeSpecName: "container-storage-root") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "container-storage-root". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.819107 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-push") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "builder-dockercfg-dkmk9-push". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.819493 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull" (OuterVolumeSpecName: "builder-dockercfg-dkmk9-pull") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "builder-dockercfg-dkmk9-pull". PluginName "kubernetes.io/secret", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.820442 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv" (OuterVolumeSpecName: "kube-api-access-vzjhv") pod "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" (UID: "8ede235c-a0a4-41df-9ed1-d8bbcbd8472a"). InnerVolumeSpecName "kube-api-access-vzjhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.913070 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-push\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-push\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.913116 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzjhv\" (UniqueName: \"kubernetes.io/projected/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-kube-api-access-vzjhv\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.913129 4730 reconciler_common.go:293] "Volume detached for volume \"build-proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-build-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.913144 4730 reconciler_common.go:293] "Volume detached for volume \"builder-dockercfg-dkmk9-pull\" (UniqueName: \"kubernetes.io/secret/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-builder-dockercfg-dkmk9-pull\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:12 crc kubenswrapper[4730]: I0221 00:18:12.913156 4730 reconciler_common.go:293] "Volume detached for volume \"container-storage-root\" (UniqueName: \"kubernetes.io/empty-dir/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a-container-storage-root\") on node \"crc\" DevicePath \"\"" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.293768 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/service-telemetry_service-telemetry-operator-5-build_8ede235c-a0a4-41df-9ed1-d8bbcbd8472a/git-clone/0.log" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.293972 4730 generic.go:334] "Generic (PLEG): container finished" podID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" containerID="79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c" exitCode=1 Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.294058 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a","Type":"ContainerDied","Data":"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c"} Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.294103 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/service-telemetry-operator-5-build" event={"ID":"8ede235c-a0a4-41df-9ed1-d8bbcbd8472a","Type":"ContainerDied","Data":"a2a88ed3fc40a76129c8f1530731ac8404ca788dfb4c2831b319858e74e78610"} Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.294063 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="service-telemetry/service-telemetry-operator-5-build" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.294194 4730 scope.go:117] "RemoveContainer" containerID="79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.342222 4730 scope.go:117] "RemoveContainer" containerID="79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.354528 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:13 crc kubenswrapper[4730]: E0221 00:18:13.354723 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c\": container with ID starting with 79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c not found: ID does not exist" containerID="79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.354793 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c"} err="failed to get container status \"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c\": rpc error: code = NotFound desc = could not find container \"79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c\": container with ID starting with 79b7d54247b2cfa146434f157ff24d61a16be9d6a7f366fc5abe2e6a9753460c not found: ID does not exist" Feb 21 00:18:13 crc kubenswrapper[4730]: I0221 00:18:13.362191 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["service-telemetry/service-telemetry-operator-5-build"] Feb 21 00:18:14 crc kubenswrapper[4730]: I0221 00:18:14.450073 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" path="/var/lib/kubelet/pods/8ede235c-a0a4-41df-9ed1-d8bbcbd8472a/volumes" Feb 21 00:18:38 crc kubenswrapper[4730]: I0221 00:18:38.214210 4730 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.317377 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-bdlb6/must-gather-xhkmt"] Feb 21 00:18:58 crc kubenswrapper[4730]: E0221 00:18:58.317996 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" containerName="git-clone" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.318008 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" containerName="git-clone" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.318115 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ede235c-a0a4-41df-9ed1-d8bbcbd8472a" containerName="git-clone" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.318648 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.348460 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bdlb6"/"openshift-service-ca.crt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.353525 4730 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-bdlb6"/"default-dockercfg-zbtzv" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.363999 4730 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-bdlb6"/"kube-root-ca.crt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.375345 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bdlb6/must-gather-xhkmt"] Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.418614 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.418704 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtkpp\" (UniqueName: \"kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.530189 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtkpp\" (UniqueName: \"kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.530247 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.530674 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.580638 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtkpp\" (UniqueName: \"kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp\") pod \"must-gather-xhkmt\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.631950 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:18:58 crc kubenswrapper[4730]: I0221 00:18:58.864239 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-bdlb6/must-gather-xhkmt"] Feb 21 00:18:59 crc kubenswrapper[4730]: I0221 00:18:59.631814 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" event={"ID":"2d549841-e11d-45ff-acac-9067ea9c4248","Type":"ContainerStarted","Data":"0a4f5932a1f56c8481e1bbae0e757936f64ee2165c535946ddaf670ff71698e2"} Feb 21 00:19:05 crc kubenswrapper[4730]: I0221 00:19:05.681964 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" event={"ID":"2d549841-e11d-45ff-acac-9067ea9c4248","Type":"ContainerStarted","Data":"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f"} Feb 21 00:19:05 crc kubenswrapper[4730]: I0221 00:19:05.775296 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:19:05 crc kubenswrapper[4730]: I0221 00:19:05.775346 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:19:06 crc kubenswrapper[4730]: I0221 00:19:06.691386 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" event={"ID":"2d549841-e11d-45ff-acac-9067ea9c4248","Type":"ContainerStarted","Data":"29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303"} Feb 21 00:19:06 crc kubenswrapper[4730]: I0221 00:19:06.709535 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" podStartSLOduration=2.344875656 podStartE2EDuration="8.709499714s" podCreationTimestamp="2026-02-21 00:18:58 +0000 UTC" firstStartedPulling="2026-02-21 00:18:58.878563229 +0000 UTC m=+791.169398959" lastFinishedPulling="2026-02-21 00:19:05.243187277 +0000 UTC m=+797.534023017" observedRunningTime="2026-02-21 00:19:06.706830297 +0000 UTC m=+798.997666037" watchObservedRunningTime="2026-02-21 00:19:06.709499714 +0000 UTC m=+799.000335454" Feb 21 00:19:35 crc kubenswrapper[4730]: I0221 00:19:35.775310 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:19:35 crc kubenswrapper[4730]: I0221 00:19:35.775962 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:19:49 crc kubenswrapper[4730]: I0221 00:19:49.907850 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-zksgf_80742fc0-505d-400a-9261-90f3e5a23183/control-plane-machine-set-operator/0.log" Feb 21 00:19:49 crc kubenswrapper[4730]: I0221 00:19:49.999374 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rmq4n_031916fc-4c68-4c3f-8f77-a95d0a59b39d/kube-rbac-proxy/0.log" Feb 21 00:19:50 crc kubenswrapper[4730]: I0221 00:19:50.035366 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rmq4n_031916fc-4c68-4c3f-8f77-a95d0a59b39d/machine-api-operator/0.log" Feb 21 00:20:02 crc kubenswrapper[4730]: I0221 00:20:02.214814 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-scgp8_3d8693e2-17ec-40dd-b2e1-ccc39aa71703/cert-manager-controller/0.log" Feb 21 00:20:02 crc kubenswrapper[4730]: I0221 00:20:02.285989 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-4vngm_1a9c9953-4d26-4895-bd21-d7e9c7f95a71/cert-manager-cainjector/0.log" Feb 21 00:20:02 crc kubenswrapper[4730]: I0221 00:20:02.370232 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-mtvcd_359d616a-dce4-4e42-bec5-b4fd29472159/cert-manager-webhook/0.log" Feb 21 00:20:05 crc kubenswrapper[4730]: I0221 00:20:05.775323 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:20:05 crc kubenswrapper[4730]: I0221 00:20:05.775844 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:20:05 crc kubenswrapper[4730]: I0221 00:20:05.775896 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:20:05 crc kubenswrapper[4730]: I0221 00:20:05.776566 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e"} pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 21 00:20:05 crc kubenswrapper[4730]: I0221 00:20:05.776620 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" containerID="cri-o://c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e" gracePeriod=600 Feb 21 00:20:06 crc kubenswrapper[4730]: I0221 00:20:06.097603 4730 generic.go:334] "Generic (PLEG): container finished" podID="33513de3-5480-4aef-87ff-879f9e7a475f" containerID="c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e" exitCode=0 Feb 21 00:20:06 crc kubenswrapper[4730]: I0221 00:20:06.097697 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerDied","Data":"c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e"} Feb 21 00:20:06 crc kubenswrapper[4730]: I0221 00:20:06.097954 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"5635ebe0d580dac1c144f6bcd57f180fa60b0727aae0d468d538ffb4df1e727e"} Feb 21 00:20:06 crc kubenswrapper[4730]: I0221 00:20:06.097982 4730 scope.go:117] "RemoveContainer" containerID="7c8cb58055875db287dd7a7e52dd7196ff3d48af4090142ec91443d86c7bb04c" Feb 21 00:20:15 crc kubenswrapper[4730]: I0221 00:20:15.888536 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-kfkqw_018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd/prometheus-operator/0.log" Feb 21 00:20:16 crc kubenswrapper[4730]: I0221 00:20:16.023462 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5_7d174177-75da-4fb4-a195-a3e06804ba1e/prometheus-operator-admission-webhook/0.log" Feb 21 00:20:16 crc kubenswrapper[4730]: I0221 00:20:16.144178 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7_5e6832e1-eb7d-4f11-9e11-95c61cbf0934/prometheus-operator-admission-webhook/0.log" Feb 21 00:20:16 crc kubenswrapper[4730]: I0221 00:20:16.310287 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-wh7ls_e7055108-18d7-4bc4-8644-a8d80b1605b3/operator/0.log" Feb 21 00:20:16 crc kubenswrapper[4730]: I0221 00:20:16.388283 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-9vq9k_b9cd1a14-f993-46fc-816b-884442179c47/perses-operator/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.055677 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/util/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.201159 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/util/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.229007 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/pull/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.262479 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/pull/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.466193 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/extract/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.468031 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/util/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.497833 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_00e596f74c7ff6aa630d3bf44b91123ebafce6c9d7df4104f82e2338f1nwtwz_e8441d0d-ff6d-4b25-9460-cf08a6f099dd/pull/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.651373 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/util/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.937862 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/pull/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.945225 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/util/0.log" Feb 21 00:20:30 crc kubenswrapper[4730]: I0221 00:20:30.974453 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/pull/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.111543 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/pull/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.182273 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/util/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.187863 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fs7cp4_adc2b228-59a9-41d8-a948-83addaa4b8f8/extract/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.352745 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/util/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.552987 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/util/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.583323 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/pull/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.605833 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/pull/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.781382 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/util/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.792607 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/pull/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.822107 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5dthss_64321551-df8b-4344-a637-b21a8ab305ea/extract/0.log" Feb 21 00:20:31 crc kubenswrapper[4730]: I0221 00:20:31.956524 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/util/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.132947 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/util/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.135516 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/pull/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.183674 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/pull/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.296544 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/util/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.349595 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/extract/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.368426 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08x2vbk_019f59b9-69c1-412c-91e7-38539454c642/pull/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.475828 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-utilities/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.752800 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-content/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.764536 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-utilities/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.794739 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-content/0.log" Feb 21 00:20:32 crc kubenswrapper[4730]: I0221 00:20:32.911653 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.040723 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/extract-content/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.107039 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lwqnm_f5ea3922-1e00-45de-9b77-881b5f8efe58/registry-server/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.153004 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.282417 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.308319 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-content/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.316053 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-content/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.467397 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.499618 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/extract-content/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.620029 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-tr8gl_308c905e-1c42-49c8-8758-67231973e7a8/registry-server/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.658329 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-rs768_6d866940-c241-40d7-b8e6-0a10dedbf9d1/marketplace-operator/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.689076 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.899359 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-content/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.899787 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-utilities/0.log" Feb 21 00:20:33 crc kubenswrapper[4730]: I0221 00:20:33.943236 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-content/0.log" Feb 21 00:20:34 crc kubenswrapper[4730]: I0221 00:20:34.081536 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-content/0.log" Feb 21 00:20:34 crc kubenswrapper[4730]: I0221 00:20:34.145315 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/extract-utilities/0.log" Feb 21 00:20:34 crc kubenswrapper[4730]: I0221 00:20:34.173533 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-g5wkr_f981a395-9367-4b4b-9ef8-d3aeee5a65f3/registry-server/0.log" Feb 21 00:20:47 crc kubenswrapper[4730]: I0221 00:20:47.334280 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-kfkqw_018ca2e3-2fff-4b8c-b5b1-3d56c9d3ebfd/prometheus-operator/0.log" Feb 21 00:20:47 crc kubenswrapper[4730]: I0221 00:20:47.337791 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-856ddc8f78-dmgt7_5e6832e1-eb7d-4f11-9e11-95c61cbf0934/prometheus-operator-admission-webhook/0.log" Feb 21 00:20:47 crc kubenswrapper[4730]: I0221 00:20:47.352471 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-856ddc8f78-cdkk5_7d174177-75da-4fb4-a195-a3e06804ba1e/prometheus-operator-admission-webhook/0.log" Feb 21 00:20:47 crc kubenswrapper[4730]: I0221 00:20:47.508730 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-wh7ls_e7055108-18d7-4bc4-8644-a8d80b1605b3/operator/0.log" Feb 21 00:20:47 crc kubenswrapper[4730]: I0221 00:20:47.541168 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-9vq9k_b9cd1a14-f993-46fc-816b-884442179c47/perses-operator/0.log" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.674964 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.681272 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.683810 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.722119 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.722257 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hwjr\" (UniqueName: \"kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.722326 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.823503 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.823587 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hwjr\" (UniqueName: \"kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.823618 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.824052 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.824136 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:24 crc kubenswrapper[4730]: I0221 00:21:24.849784 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hwjr\" (UniqueName: \"kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr\") pod \"redhat-operators-xs45m\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:25 crc kubenswrapper[4730]: I0221 00:21:25.000057 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:25 crc kubenswrapper[4730]: I0221 00:21:25.196744 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:25 crc kubenswrapper[4730]: I0221 00:21:25.620148 4730 generic.go:334] "Generic (PLEG): container finished" podID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerID="7b0e9301bc8ff32dca0e0b498ff418cec96aeee20d73f5fec1244e3ba47c63a9" exitCode=0 Feb 21 00:21:25 crc kubenswrapper[4730]: I0221 00:21:25.620208 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerDied","Data":"7b0e9301bc8ff32dca0e0b498ff418cec96aeee20d73f5fec1244e3ba47c63a9"} Feb 21 00:21:25 crc kubenswrapper[4730]: I0221 00:21:25.620246 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerStarted","Data":"7cb713c59173160f414cd7d123c01ba47d92da9a5af628acee3121acc06cb671"} Feb 21 00:21:26 crc kubenswrapper[4730]: I0221 00:21:26.661669 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerStarted","Data":"0ec0239447b6ce24509e1e1f674c71830ba7fe6cfe8c54a114f49fe7e0775856"} Feb 21 00:21:27 crc kubenswrapper[4730]: I0221 00:21:27.668683 4730 generic.go:334] "Generic (PLEG): container finished" podID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerID="0ec0239447b6ce24509e1e1f674c71830ba7fe6cfe8c54a114f49fe7e0775856" exitCode=0 Feb 21 00:21:27 crc kubenswrapper[4730]: I0221 00:21:27.668721 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerDied","Data":"0ec0239447b6ce24509e1e1f674c71830ba7fe6cfe8c54a114f49fe7e0775856"} Feb 21 00:21:28 crc kubenswrapper[4730]: I0221 00:21:28.680168 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerStarted","Data":"b157ea9d8a26da85d65f1ca0e608c144303b810021abc5b923b954a729acfca5"} Feb 21 00:21:28 crc kubenswrapper[4730]: I0221 00:21:28.704057 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xs45m" podStartSLOduration=2.265985496 podStartE2EDuration="4.704033765s" podCreationTimestamp="2026-02-21 00:21:24 +0000 UTC" firstStartedPulling="2026-02-21 00:21:25.621981565 +0000 UTC m=+937.912817315" lastFinishedPulling="2026-02-21 00:21:28.060029834 +0000 UTC m=+940.350865584" observedRunningTime="2026-02-21 00:21:28.697259729 +0000 UTC m=+940.988095469" watchObservedRunningTime="2026-02-21 00:21:28.704033765 +0000 UTC m=+940.994869495" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.058956 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.061199 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.079741 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.214929 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.214999 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.215104 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zjpq\" (UniqueName: \"kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.316629 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.317011 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.317505 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.317532 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zjpq\" (UniqueName: \"kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.317741 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.338566 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zjpq\" (UniqueName: \"kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq\") pod \"certified-operators-4jcx2\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.416435 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.679969 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:31 crc kubenswrapper[4730]: I0221 00:21:31.704645 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerStarted","Data":"dcbd6544fe3e26ba5299c4a57638f3718ebff3f9e7d8699858856a0bb33815d4"} Feb 21 00:21:32 crc kubenswrapper[4730]: I0221 00:21:32.714523 4730 generic.go:334] "Generic (PLEG): container finished" podID="e1396d49-2129-4483-b03e-ec9976824c44" containerID="ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402" exitCode=0 Feb 21 00:21:32 crc kubenswrapper[4730]: I0221 00:21:32.715052 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerDied","Data":"ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402"} Feb 21 00:21:32 crc kubenswrapper[4730]: I0221 00:21:32.721048 4730 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Feb 21 00:21:33 crc kubenswrapper[4730]: I0221 00:21:33.722353 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerStarted","Data":"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e"} Feb 21 00:21:34 crc kubenswrapper[4730]: I0221 00:21:34.730704 4730 generic.go:334] "Generic (PLEG): container finished" podID="e1396d49-2129-4483-b03e-ec9976824c44" containerID="9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e" exitCode=0 Feb 21 00:21:34 crc kubenswrapper[4730]: I0221 00:21:34.730769 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerDied","Data":"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e"} Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.009573 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.009624 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.059764 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.739458 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerStarted","Data":"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500"} Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.762584 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4jcx2" podStartSLOduration=2.367908551 podStartE2EDuration="4.762564774s" podCreationTimestamp="2026-02-21 00:21:31 +0000 UTC" firstStartedPulling="2026-02-21 00:21:32.720358007 +0000 UTC m=+945.011193777" lastFinishedPulling="2026-02-21 00:21:35.11501426 +0000 UTC m=+947.405850000" observedRunningTime="2026-02-21 00:21:35.75760684 +0000 UTC m=+948.048442580" watchObservedRunningTime="2026-02-21 00:21:35.762564774 +0000 UTC m=+948.053400504" Feb 21 00:21:35 crc kubenswrapper[4730]: I0221 00:21:35.788508 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:37 crc kubenswrapper[4730]: I0221 00:21:37.441011 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:37 crc kubenswrapper[4730]: I0221 00:21:37.755384 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xs45m" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="registry-server" containerID="cri-o://b157ea9d8a26da85d65f1ca0e608c144303b810021abc5b923b954a729acfca5" gracePeriod=2 Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.777548 4730 generic.go:334] "Generic (PLEG): container finished" podID="2d549841-e11d-45ff-acac-9067ea9c4248" containerID="416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f" exitCode=0 Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.777862 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" event={"ID":"2d549841-e11d-45ff-acac-9067ea9c4248","Type":"ContainerDied","Data":"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f"} Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.778460 4730 scope.go:117] "RemoveContainer" containerID="416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f" Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.781638 4730 generic.go:334] "Generic (PLEG): container finished" podID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerID="b157ea9d8a26da85d65f1ca0e608c144303b810021abc5b923b954a729acfca5" exitCode=0 Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.781664 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerDied","Data":"b157ea9d8a26da85d65f1ca0e608c144303b810021abc5b923b954a729acfca5"} Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.855855 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.961960 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities\") pod \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.962155 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6hwjr\" (UniqueName: \"kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr\") pod \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.962210 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content\") pod \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\" (UID: \"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7\") " Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.973954 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities" (OuterVolumeSpecName: "utilities") pod "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" (UID: "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:21:40 crc kubenswrapper[4730]: I0221 00:21:40.980141 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr" (OuterVolumeSpecName: "kube-api-access-6hwjr") pod "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" (UID: "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7"). InnerVolumeSpecName "kube-api-access-6hwjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.064552 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.064621 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6hwjr\" (UniqueName: \"kubernetes.io/projected/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-kube-api-access-6hwjr\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.120165 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" (UID: "94abf3cc-9c65-45bd-9a01-9c5745ea8ec7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.166277 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.417473 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.417628 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.474425 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.594648 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bdlb6_must-gather-xhkmt_2d549841-e11d-45ff-acac-9067ea9c4248/gather/0.log" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.791280 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xs45m" event={"ID":"94abf3cc-9c65-45bd-9a01-9c5745ea8ec7","Type":"ContainerDied","Data":"7cb713c59173160f414cd7d123c01ba47d92da9a5af628acee3121acc06cb671"} Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.791352 4730 scope.go:117] "RemoveContainer" containerID="b157ea9d8a26da85d65f1ca0e608c144303b810021abc5b923b954a729acfca5" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.791309 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xs45m" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.815016 4730 scope.go:117] "RemoveContainer" containerID="0ec0239447b6ce24509e1e1f674c71830ba7fe6cfe8c54a114f49fe7e0775856" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.820410 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.827864 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xs45m"] Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.840360 4730 scope.go:117] "RemoveContainer" containerID="7b0e9301bc8ff32dca0e0b498ff418cec96aeee20d73f5fec1244e3ba47c63a9" Feb 21 00:21:41 crc kubenswrapper[4730]: I0221 00:21:41.844308 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:42 crc kubenswrapper[4730]: I0221 00:21:42.445574 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" path="/var/lib/kubelet/pods/94abf3cc-9c65-45bd-9a01-9c5745ea8ec7/volumes" Feb 21 00:21:43 crc kubenswrapper[4730]: I0221 00:21:43.843651 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:44 crc kubenswrapper[4730]: I0221 00:21:44.815740 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4jcx2" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="registry-server" containerID="cri-o://798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500" gracePeriod=2 Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.245515 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.428973 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities\") pod \"e1396d49-2129-4483-b03e-ec9976824c44\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.429257 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content\") pod \"e1396d49-2129-4483-b03e-ec9976824c44\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.429343 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zjpq\" (UniqueName: \"kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq\") pod \"e1396d49-2129-4483-b03e-ec9976824c44\" (UID: \"e1396d49-2129-4483-b03e-ec9976824c44\") " Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.430002 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities" (OuterVolumeSpecName: "utilities") pod "e1396d49-2129-4483-b03e-ec9976824c44" (UID: "e1396d49-2129-4483-b03e-ec9976824c44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.437988 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq" (OuterVolumeSpecName: "kube-api-access-8zjpq") pod "e1396d49-2129-4483-b03e-ec9976824c44" (UID: "e1396d49-2129-4483-b03e-ec9976824c44"). InnerVolumeSpecName "kube-api-access-8zjpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.501062 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e1396d49-2129-4483-b03e-ec9976824c44" (UID: "e1396d49-2129-4483-b03e-ec9976824c44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.530588 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.530629 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zjpq\" (UniqueName: \"kubernetes.io/projected/e1396d49-2129-4483-b03e-ec9976824c44-kube-api-access-8zjpq\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.530646 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e1396d49-2129-4483-b03e-ec9976824c44-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.836330 4730 generic.go:334] "Generic (PLEG): container finished" podID="e1396d49-2129-4483-b03e-ec9976824c44" containerID="798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500" exitCode=0 Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.836393 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerDied","Data":"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500"} Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.836431 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4jcx2" event={"ID":"e1396d49-2129-4483-b03e-ec9976824c44","Type":"ContainerDied","Data":"dcbd6544fe3e26ba5299c4a57638f3718ebff3f9e7d8699858856a0bb33815d4"} Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.836455 4730 scope.go:117] "RemoveContainer" containerID="798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.836498 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4jcx2" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.861921 4730 scope.go:117] "RemoveContainer" containerID="9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.889864 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.901440 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4jcx2"] Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.906228 4730 scope.go:117] "RemoveContainer" containerID="ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.929276 4730 scope.go:117] "RemoveContainer" containerID="798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500" Feb 21 00:21:45 crc kubenswrapper[4730]: E0221 00:21:45.929837 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500\": container with ID starting with 798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500 not found: ID does not exist" containerID="798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.929894 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500"} err="failed to get container status \"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500\": rpc error: code = NotFound desc = could not find container \"798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500\": container with ID starting with 798d6b13f86f1b361ce95e8bf016e6ac80c844809bc736568c3ad7ddeb6ac500 not found: ID does not exist" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.929948 4730 scope.go:117] "RemoveContainer" containerID="9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e" Feb 21 00:21:45 crc kubenswrapper[4730]: E0221 00:21:45.930350 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e\": container with ID starting with 9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e not found: ID does not exist" containerID="9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.930502 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e"} err="failed to get container status \"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e\": rpc error: code = NotFound desc = could not find container \"9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e\": container with ID starting with 9b825f7b4e99526163798ce23fb033c5f19c72ed34273963ba8a1661fb481e8e not found: ID does not exist" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.930622 4730 scope.go:117] "RemoveContainer" containerID="ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402" Feb 21 00:21:45 crc kubenswrapper[4730]: E0221 00:21:45.931059 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402\": container with ID starting with ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402 not found: ID does not exist" containerID="ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402" Feb 21 00:21:45 crc kubenswrapper[4730]: I0221 00:21:45.931092 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402"} err="failed to get container status \"ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402\": rpc error: code = NotFound desc = could not find container \"ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402\": container with ID starting with ff195b280af142713b0c31622af57f11188c170a072c174692eef92199283402 not found: ID does not exist" Feb 21 00:21:46 crc kubenswrapper[4730]: I0221 00:21:46.444949 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1396d49-2129-4483-b03e-ec9976824c44" path="/var/lib/kubelet/pods/e1396d49-2129-4483-b03e-ec9976824c44/volumes" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.238010 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-bdlb6/must-gather-xhkmt"] Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.238548 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="copy" containerID="cri-o://29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303" gracePeriod=2 Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.243392 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-bdlb6/must-gather-xhkmt"] Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.613211 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bdlb6_must-gather-xhkmt_2d549841-e11d-45ff-acac-9067ea9c4248/copy/0.log" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.613739 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.683478 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtkpp\" (UniqueName: \"kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp\") pod \"2d549841-e11d-45ff-acac-9067ea9c4248\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.683586 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output\") pod \"2d549841-e11d-45ff-acac-9067ea9c4248\" (UID: \"2d549841-e11d-45ff-acac-9067ea9c4248\") " Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.688833 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp" (OuterVolumeSpecName: "kube-api-access-dtkpp") pod "2d549841-e11d-45ff-acac-9067ea9c4248" (UID: "2d549841-e11d-45ff-acac-9067ea9c4248"). InnerVolumeSpecName "kube-api-access-dtkpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.730318 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "2d549841-e11d-45ff-acac-9067ea9c4248" (UID: "2d549841-e11d-45ff-acac-9067ea9c4248"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.784783 4730 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/2d549841-e11d-45ff-acac-9067ea9c4248-must-gather-output\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.784821 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtkpp\" (UniqueName: \"kubernetes.io/projected/2d549841-e11d-45ff-acac-9067ea9c4248-kube-api-access-dtkpp\") on node \"crc\" DevicePath \"\"" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.860475 4730 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-bdlb6_must-gather-xhkmt_2d549841-e11d-45ff-acac-9067ea9c4248/copy/0.log" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.860817 4730 generic.go:334] "Generic (PLEG): container finished" podID="2d549841-e11d-45ff-acac-9067ea9c4248" containerID="29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303" exitCode=143 Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.860913 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-bdlb6/must-gather-xhkmt" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.860936 4730 scope.go:117] "RemoveContainer" containerID="29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.895168 4730 scope.go:117] "RemoveContainer" containerID="416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.941927 4730 scope.go:117] "RemoveContainer" containerID="29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303" Feb 21 00:21:48 crc kubenswrapper[4730]: E0221 00:21:48.942355 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303\": container with ID starting with 29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303 not found: ID does not exist" containerID="29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.942389 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303"} err="failed to get container status \"29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303\": rpc error: code = NotFound desc = could not find container \"29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303\": container with ID starting with 29a9168bec6d5621712e74a23ad9ac5ca7991da0b466ca355b4eaa25f6ad3303 not found: ID does not exist" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.942414 4730 scope.go:117] "RemoveContainer" containerID="416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f" Feb 21 00:21:48 crc kubenswrapper[4730]: E0221 00:21:48.942750 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f\": container with ID starting with 416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f not found: ID does not exist" containerID="416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f" Feb 21 00:21:48 crc kubenswrapper[4730]: I0221 00:21:48.942779 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f"} err="failed to get container status \"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f\": rpc error: code = NotFound desc = could not find container \"416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f\": container with ID starting with 416b160ced257aed718171289f589a994ecea62a9486c5ccd40498706411e19f not found: ID does not exist" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.048890 4730 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049158 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="extract-utilities" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049177 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="extract-utilities" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049192 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="extract-content" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049199 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="extract-content" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049215 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="extract-content" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049222 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="extract-content" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049233 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="copy" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049240 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="copy" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049250 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="extract-utilities" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049257 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="extract-utilities" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049265 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049272 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049283 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049290 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: E0221 00:21:49.049312 4730 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="gather" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049319 4730 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="gather" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049426 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="94abf3cc-9c65-45bd-9a01-9c5745ea8ec7" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049439 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="copy" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049453 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" containerName="gather" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.049465 4730 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1396d49-2129-4483-b03e-ec9976824c44" containerName="registry-server" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.050410 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.060538 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.088970 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.089019 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.089082 4730 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjtzw\" (UniqueName: \"kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.190771 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.190820 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.190843 4730 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjtzw\" (UniqueName: \"kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.191329 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.191478 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.216960 4730 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjtzw\" (UniqueName: \"kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw\") pod \"community-operators-kgrkp\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.366710 4730 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.650710 4730 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.868740 4730 generic.go:334] "Generic (PLEG): container finished" podID="ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" containerID="73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29" exitCode=0 Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.868837 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerDied","Data":"73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29"} Feb 21 00:21:49 crc kubenswrapper[4730]: I0221 00:21:49.869111 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerStarted","Data":"d11975a716501a9a3bcde6cfc4396cd1ed1296d6d3d30ac7fe6d51899aa12f67"} Feb 21 00:21:50 crc kubenswrapper[4730]: I0221 00:21:50.445498 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d549841-e11d-45ff-acac-9067ea9c4248" path="/var/lib/kubelet/pods/2d549841-e11d-45ff-acac-9067ea9c4248/volumes" Feb 21 00:21:50 crc kubenswrapper[4730]: I0221 00:21:50.876611 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerStarted","Data":"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444"} Feb 21 00:21:51 crc kubenswrapper[4730]: E0221 00:21:51.035169 4730 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca6c0fc3_99b6_439c_bbfd_aae0af30ff40.slice/crio-9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca6c0fc3_99b6_439c_bbfd_aae0af30ff40.slice/crio-conmon-9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444.scope\": RecentStats: unable to find data in memory cache]" Feb 21 00:21:51 crc kubenswrapper[4730]: I0221 00:21:51.884285 4730 generic.go:334] "Generic (PLEG): container finished" podID="ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" containerID="9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444" exitCode=0 Feb 21 00:21:51 crc kubenswrapper[4730]: I0221 00:21:51.884358 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerDied","Data":"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444"} Feb 21 00:21:53 crc kubenswrapper[4730]: I0221 00:21:53.896199 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerStarted","Data":"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582"} Feb 21 00:21:59 crc kubenswrapper[4730]: I0221 00:21:59.367525 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:59 crc kubenswrapper[4730]: I0221 00:21:59.368237 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:59 crc kubenswrapper[4730]: I0221 00:21:59.418505 4730 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:21:59 crc kubenswrapper[4730]: I0221 00:21:59.442967 4730 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-kgrkp" podStartSLOduration=6.939240423 podStartE2EDuration="10.442951098s" podCreationTimestamp="2026-02-21 00:21:49 +0000 UTC" firstStartedPulling="2026-02-21 00:21:49.870373681 +0000 UTC m=+962.161209421" lastFinishedPulling="2026-02-21 00:21:53.374084366 +0000 UTC m=+965.664920096" observedRunningTime="2026-02-21 00:21:53.937342349 +0000 UTC m=+966.228178079" watchObservedRunningTime="2026-02-21 00:21:59.442951098 +0000 UTC m=+971.733786828" Feb 21 00:21:59 crc kubenswrapper[4730]: I0221 00:21:59.976672 4730 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:22:00 crc kubenswrapper[4730]: I0221 00:22:00.027009 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:22:01 crc kubenswrapper[4730]: I0221 00:22:01.984385 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-kgrkp" podUID="ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" containerName="registry-server" containerID="cri-o://edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582" gracePeriod=2 Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.441100 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.565885 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content\") pod \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.565970 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjtzw\" (UniqueName: \"kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw\") pod \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.566813 4730 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities\") pod \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\" (UID: \"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40\") " Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.567580 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities" (OuterVolumeSpecName: "utilities") pod "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" (UID: "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.571731 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw" (OuterVolumeSpecName: "kube-api-access-cjtzw") pod "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" (UID: "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40"). InnerVolumeSpecName "kube-api-access-cjtzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.625478 4730 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" (UID: "ca6c0fc3-99b6-439c-bbfd-aae0af30ff40"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.667652 4730 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-catalog-content\") on node \"crc\" DevicePath \"\"" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.667689 4730 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjtzw\" (UniqueName: \"kubernetes.io/projected/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-kube-api-access-cjtzw\") on node \"crc\" DevicePath \"\"" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.667704 4730 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40-utilities\") on node \"crc\" DevicePath \"\"" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.991585 4730 generic.go:334] "Generic (PLEG): container finished" podID="ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" containerID="edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582" exitCode=0 Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.991626 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerDied","Data":"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582"} Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.991661 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-kgrkp" event={"ID":"ca6c0fc3-99b6-439c-bbfd-aae0af30ff40","Type":"ContainerDied","Data":"d11975a716501a9a3bcde6cfc4396cd1ed1296d6d3d30ac7fe6d51899aa12f67"} Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.991679 4730 scope.go:117] "RemoveContainer" containerID="edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582" Feb 21 00:22:02 crc kubenswrapper[4730]: I0221 00:22:02.991695 4730 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-kgrkp" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.015288 4730 scope.go:117] "RemoveContainer" containerID="9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.036654 4730 scope.go:117] "RemoveContainer" containerID="73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.067827 4730 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.072323 4730 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-kgrkp"] Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.075632 4730 scope.go:117] "RemoveContainer" containerID="edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582" Feb 21 00:22:03 crc kubenswrapper[4730]: E0221 00:22:03.076217 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582\": container with ID starting with edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582 not found: ID does not exist" containerID="edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.076263 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582"} err="failed to get container status \"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582\": rpc error: code = NotFound desc = could not find container \"edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582\": container with ID starting with edc336fe0a769d413593b936bd4bdd8790eb37421c821fce4e49c19310084582 not found: ID does not exist" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.076301 4730 scope.go:117] "RemoveContainer" containerID="9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444" Feb 21 00:22:03 crc kubenswrapper[4730]: E0221 00:22:03.076719 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444\": container with ID starting with 9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444 not found: ID does not exist" containerID="9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.076780 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444"} err="failed to get container status \"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444\": rpc error: code = NotFound desc = could not find container \"9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444\": container with ID starting with 9ac56dedb8d370b7bf0860afa1ce7972b46355c6ff43cdd6560ee96c667aa444 not found: ID does not exist" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.076885 4730 scope.go:117] "RemoveContainer" containerID="73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29" Feb 21 00:22:03 crc kubenswrapper[4730]: E0221 00:22:03.077465 4730 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29\": container with ID starting with 73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29 not found: ID does not exist" containerID="73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29" Feb 21 00:22:03 crc kubenswrapper[4730]: I0221 00:22:03.077488 4730 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29"} err="failed to get container status \"73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29\": rpc error: code = NotFound desc = could not find container \"73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29\": container with ID starting with 73463cff5f3bcbf1c6b89e37968acdab2d689d666a5dfb33b57481a8e0cb4c29 not found: ID does not exist" Feb 21 00:22:04 crc kubenswrapper[4730]: I0221 00:22:04.444140 4730 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca6c0fc3-99b6-439c-bbfd-aae0af30ff40" path="/var/lib/kubelet/pods/ca6c0fc3-99b6-439c-bbfd-aae0af30ff40/volumes" Feb 21 00:22:35 crc kubenswrapper[4730]: I0221 00:22:35.774913 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:22:35 crc kubenswrapper[4730]: I0221 00:22:35.775487 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:23:05 crc kubenswrapper[4730]: I0221 00:23:05.776074 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:23:05 crc kubenswrapper[4730]: I0221 00:23:05.776763 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:23:35 crc kubenswrapper[4730]: I0221 00:23:35.775894 4730 patch_prober.go:28] interesting pod/machine-config-daemon-r8q7m container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Feb 21 00:23:35 crc kubenswrapper[4730]: I0221 00:23:35.777728 4730 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Feb 21 00:23:35 crc kubenswrapper[4730]: I0221 00:23:35.777877 4730 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" Feb 21 00:23:35 crc kubenswrapper[4730]: I0221 00:23:35.778597 4730 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5635ebe0d580dac1c144f6bcd57f180fa60b0727aae0d468d538ffb4df1e727e"} pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Feb 21 00:23:35 crc kubenswrapper[4730]: I0221 00:23:35.778793 4730 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" podUID="33513de3-5480-4aef-87ff-879f9e7a475f" containerName="machine-config-daemon" containerID="cri-o://5635ebe0d580dac1c144f6bcd57f180fa60b0727aae0d468d538ffb4df1e727e" gracePeriod=600 Feb 21 00:23:36 crc kubenswrapper[4730]: I0221 00:23:36.694368 4730 generic.go:334] "Generic (PLEG): container finished" podID="33513de3-5480-4aef-87ff-879f9e7a475f" containerID="5635ebe0d580dac1c144f6bcd57f180fa60b0727aae0d468d538ffb4df1e727e" exitCode=0 Feb 21 00:23:36 crc kubenswrapper[4730]: I0221 00:23:36.694418 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerDied","Data":"5635ebe0d580dac1c144f6bcd57f180fa60b0727aae0d468d538ffb4df1e727e"} Feb 21 00:23:36 crc kubenswrapper[4730]: I0221 00:23:36.695075 4730 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-r8q7m" event={"ID":"33513de3-5480-4aef-87ff-879f9e7a475f","Type":"ContainerStarted","Data":"54d890f24e67f46be1687c18489ed5296a34065a310bc4333a6ed72a895d525d"} Feb 21 00:23:36 crc kubenswrapper[4730]: I0221 00:23:36.695115 4730 scope.go:117] "RemoveContainer" containerID="c4416e95cff771d010e3ee0e0f003a67e7a54468c55add51d6f75c837bdb801e" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515146175557024465 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015146175560017374 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015146173023016507 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015146173023015457 5ustar corecore